diff --git a/cts/cli/regression.crm_mon.exp b/cts/cli/regression.crm_mon.exp
index 070a3205a3..a45037aeea 100644
--- a/cts/cli/regression.crm_mon.exp
+++ b/cts/cli/regression.crm_mon.exp
@@ -1,3966 +1,3966 @@
=#=#=#= Begin test: Basic text output =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
* Fencing (stonith:fence_xvm): Started cluster01
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
* Unpromoted: [ cluster01 ]
=#=#=#= End test: Basic text output - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output
=#=#=#= Begin test: XML output =#=#=#=
=#=#=#= End test: XML output - OK (0) =#=#=#=
* Passed: crm_mon - XML output
=#=#=#= Begin test: Basic text output without node section =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
* Fencing (stonith:fence_xvm): Started cluster01
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
* Unpromoted: [ cluster01 ]
=#=#=#= End test: Basic text output without node section - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output without node section
=#=#=#= Begin test: XML output without the node section =#=#=#=
=#=#=#= End test: XML output without the node section - OK (0) =#=#=#=
* Passed: crm_mon - XML output without the node section
=#=#=#= Begin test: Text output with only the node section =#=#=#=
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
=#=#=#= End test: Text output with only the node section - OK (0) =#=#=#=
* Passed: crm_mon - Text output with only the node section
=#=#=#= Begin test: Complete text output =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
* Fencing (stonith:fence_xvm): Started cluster01
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
* Unpromoted: [ cluster01 ]
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster01:
* httpd: migration-threshold=1000000:
* (1) start
* Node: httpd-bundle-1@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete text output - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output
=#=#=#= Begin test: Complete text output with detail =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* ping (ocf:pacemaker:ping): Started cluster02
* ping (ocf:pacemaker:ping): Started cluster01
* Fencing (stonith:fence_xvm): Started cluster01
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Container bundle set: httpd-bundle [pcmk:http]:
* Replica[0]
* httpd-bundle-ip-192.168.122.131 (ocf:heartbeat:IPaddr2): Started cluster01
* httpd (ocf:heartbeat:apache): Started httpd-bundle-0
* httpd-bundle-docker-0 (ocf:heartbeat:docker): Started cluster01
* httpd-bundle-0 (ocf:pacemaker:remote): Started cluster01
* Replica[1]
* httpd-bundle-ip-192.168.122.132 (ocf:heartbeat:IPaddr2): Started cluster02
* httpd (ocf:heartbeat:apache): Started httpd-bundle-1
* httpd-bundle-docker-1 (ocf:heartbeat:docker): Started cluster02
* httpd-bundle-1 (ocf:pacemaker:remote): Started cluster02
* Replica[2]
* httpd-bundle-ip-192.168.122.133 (ocf:heartbeat:IPaddr2): Stopped
* httpd (ocf:heartbeat:apache): Stopped
* httpd-bundle-docker-2 (ocf:heartbeat:docker): Stopped
* httpd-bundle-2 (ocf:pacemaker:remote): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:0:
* mysql-proxy (lsb:mysql-proxy): Started cluster02
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* promotable-rsc (ocf:pacemaker:Stateful): Promoted cluster02
* promotable-rsc (ocf:pacemaker:Stateful): Unpromoted cluster01
* promotable-rsc (ocf:pacemaker:Stateful): Stopped
* promotable-rsc (ocf:pacemaker:Stateful): Stopped
* promotable-rsc (ocf:pacemaker:Stateful): Stopped
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: cluster01 (1):
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster01:
* httpd: migration-threshold=1000000:
* (1) start
* Node: httpd-bundle-1@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01 (1)
=#=#=#= End test: Complete text output with detail - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output with detail
=#=#=#= Begin test: Complete brief text output =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* 1 (ocf:pacemaker:Dummy): Active cluster02
* 1 (stonith:fence_xvm): Active cluster01
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* 1/1 (lsb:exim): Active cluster02
* 1/1 (ocf:heartbeat:IPaddr): Active cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
* Unpromoted: [ cluster01 ]
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster01:
* httpd: migration-threshold=1000000:
* (1) start
* Node: httpd-bundle-1@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete brief text output - OK (0) =#=#=#=
* Passed: crm_mon - Complete brief text output
=#=#=#= Begin test: Complete text output grouped by node =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Node cluster01: online:
* Resources:
* ping (ocf:pacemaker:ping): Started
* Fencing (stonith:fence_xvm): Started
* mysql-proxy (lsb:mysql-proxy): Started
* promotable-rsc (ocf:pacemaker:Stateful): Unpromoted
* httpd-bundle-ip-192.168.122.131 (ocf:heartbeat:IPaddr2): Started
* httpd-bundle-docker-0 (ocf:heartbeat:docker): Started
* Node cluster02: online:
* Resources:
* ping (ocf:pacemaker:ping): Started
* dummy (ocf:pacemaker:Dummy): Started
* Public-IP (ocf:heartbeat:IPaddr): Started
* Email (lsb:exim): Started
* mysql-proxy (lsb:mysql-proxy): Started
* promotable-rsc (ocf:pacemaker:Stateful): Promoted
* httpd-bundle-ip-192.168.122.132 (ocf:heartbeat:IPaddr2): Started
* httpd-bundle-docker-1 (ocf:heartbeat:docker): Started
* GuestNode httpd-bundle-0@cluster01: online:
* Resources:
* httpd (ocf:heartbeat:apache): Started
* GuestNode httpd-bundle-1@cluster02: online:
* Resources:
* httpd (ocf:heartbeat:apache): Started
* GuestNode httpd-bundle-2@: OFFLINE:
* Resources:
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster01:
* httpd: migration-threshold=1000000:
* (1) start
* Node: httpd-bundle-1@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete text output grouped by node - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output grouped by node
=#=#=#= Begin test: Complete brief text output grouped by node =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Node cluster01: online:
* Resources:
* 1 (lsb:mysql-proxy): Active
* 1 (ocf:heartbeat:IPaddr2): Active
* 1 (ocf:heartbeat:docker): Active
* 1 (ocf:pacemaker:Stateful): Active
* 1 (ocf:pacemaker:ping): Active
* 1 (ocf:pacemaker:remote): Active
* 1 (stonith:fence_xvm): Active
* Node cluster02: online:
* Resources:
* 1 (lsb:exim): Active
* 1 (lsb:mysql-proxy): Active
* 1 (ocf:heartbeat:IPaddr): Active
* 1 (ocf:heartbeat:IPaddr2): Active
* 1 (ocf:heartbeat:docker): Active
* 1 (ocf:pacemaker:Dummy): Active
* 1 (ocf:pacemaker:Stateful): Active
* 1 (ocf:pacemaker:ping): Active
* 1 (ocf:pacemaker:remote): Active
* GuestNode httpd-bundle-0@cluster01: online:
* Resources:
* 1 (ocf:heartbeat:apache): Active
* GuestNode httpd-bundle-1@cluster02: online:
* Resources:
* 1 (ocf:heartbeat:apache): Active
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster01:
* httpd: migration-threshold=1000000:
* (1) start
* Node: httpd-bundle-1@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete brief text output grouped by node - OK (0) =#=#=#=
* Passed: crm_mon - Complete brief text output grouped by node
=#=#=#= Begin test: XML output grouped by node =#=#=#=
=#=#=#= End test: XML output grouped by node - OK (0) =#=#=#=
* Passed: crm_mon - XML output grouped by node
=#=#=#= Begin test: Complete text output filtered by node =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Fencing (stonith:fence_xvm): Started cluster01
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Unpromoted: [ cluster01 ]
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
Operations:
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* dummy: migration-threshold=1000000:
* (16) stop
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (2) start
* (4) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete text output filtered by node - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by node
=#=#=#= Begin test: XML output filtered by node =#=#=#=
=#=#=#= End test: XML output filtered by node - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by node
=#=#=#= Begin test: Complete text output filtered by tag =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster02 ]
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
Node Attributes:
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* dummy: migration-threshold=1000000:
* (18) start
* (19) monitor: interval="60000ms"
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* promotable-rsc: migration-threshold=1000000:
* (4) monitor: interval="10000ms"
* (5) cancel: interval="10000ms"
* (6) promote
* (7) monitor: interval="5000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
Negative Location Constraints:
* not-on-cluster1 prevents dummy from running on cluster01
=#=#=#= End test: Complete text output filtered by tag - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by tag
=#=#=#= Begin test: XML output filtered by tag =#=#=#=
=#=#=#= End test: XML output filtered by tag - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by tag
=#=#=#= Begin test: Complete text output filtered by resource tag =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Fencing (stonith:fence_xvm): Started cluster01
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster01:
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
=#=#=#= End test: Complete text output filtered by resource tag - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by resource tag
=#=#=#= Begin test: XML output filtered by resource tag =#=#=#=
=#=#=#= End test: XML output filtered by resource tag - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by resource tag
=#=#=#= Begin test: Basic text output filtered by node that doesn't exist =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Active Resources:
* No active resources
=#=#=#= End test: Basic text output filtered by node that doesn't exist - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output filtered by node that doesn't exist
=#=#=#= Begin test: XML output filtered by node that doesn't exist =#=#=#=
=#=#=#= End test: XML output filtered by node that doesn't exist - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by node that doesn't exist
=#=#=#= Begin test: Basic text output with inactive resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
* Fencing (stonith:fence_xvm): Started cluster01
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Clone Set: inactive-clone [inactive-dhcpd] (disabled):
* Stopped (disabled): [ cluster01 cluster02 ]
* Resource Group: inactive-group (disabled):
* inactive-dummy-1 (ocf:pacemaker:Dummy): Stopped (disabled)
* inactive-dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster01 cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
* Unpromoted: [ cluster01 ]
=#=#=#= End test: Basic text output with inactive resources - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources
=#=#=#= Begin test: Basic text output with inactive resources, filtered by node =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster02 ]
Full List of Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster02 ]
* dummy (ocf:pacemaker:Dummy): Started cluster02
* Clone Set: inactive-clone [inactive-dhcpd] (disabled):
* Stopped (disabled): [ cluster02 ]
* Resource Group: inactive-group (disabled):
* inactive-dummy-1 (ocf:pacemaker:Dummy): Stopped (disabled)
* inactive-dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
* Clone Set: mysql-clone-group [mysql-group]:
* Started: [ cluster02 ]
* Clone Set: promotable-clone [promotable-rsc] (promotable):
* Promoted: [ cluster02 ]
=#=#=#= End test: Basic text output with inactive resources, filtered by node - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by node
=#=#=#= Begin test: Complete text output filtered by primitive resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Fencing (stonith:fence_xvm): Started cluster01
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster01:
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
=#=#=#= End test: Complete text output filtered by primitive resource - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by primitive resource
=#=#=#= Begin test: XML output filtered by primitive resource =#=#=#=
=#=#=#= End test: XML output filtered by primitive resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by primitive resource
=#=#=#= Begin test: Complete text output filtered by group resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
* Email (lsb:exim): Started cluster02
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* Public-IP: migration-threshold=1000000:
* (2) start
* Email: migration-threshold=1000000:
* (2) start
=#=#=#= End test: Complete text output filtered by group resource - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by group resource
=#=#=#= Begin test: XML output filtered by group resource =#=#=#=
=#=#=#= End test: XML output filtered by group resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by group resource
=#=#=#= Begin test: Complete text output filtered by group resource member =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Resource Group: exim-group:
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* Public-IP: migration-threshold=1000000:
* (2) start
=#=#=#= End test: Complete text output filtered by group resource member - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by group resource member
=#=#=#= Begin test: XML output filtered by group resource member =#=#=#=
=#=#=#= End test: XML output filtered by group resource member - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by group resource member
=#=#=#= Begin test: Complete text output filtered by clone resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
=#=#=#= End test: Complete text output filtered by clone resource - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by clone resource
=#=#=#= Begin test: XML output filtered by clone resource =#=#=#=
=#=#=#= End test: XML output filtered by clone resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by clone resource
=#=#=#= Begin test: Complete text output filtered by clone resource instance =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 cluster02 ]
Node Attributes:
* Node: cluster01:
* location : office
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* Node: cluster01:
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
=#=#=#= End test: Complete text output filtered by clone resource instance - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by clone resource instance
=#=#=#= Begin test: XML output filtered by clone resource instance =#=#=#=
=#=#=#= End test: XML output filtered by clone resource instance - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by clone resource instance
=#=#=#= Begin test: Complete text output filtered by exact clone resource instance =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* ping (ocf:pacemaker:ping): Started cluster02
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* ping: migration-threshold=1000000:
* (11) start
* (12) monitor: interval="10000ms"
* Node: cluster01 (1):
* ping: migration-threshold=1000000:
* (17) start
* (18) monitor: interval="10000ms"
=#=#=#= End test: Complete text output filtered by exact clone resource instance - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output filtered by exact clone resource instance
=#=#=#= Begin test: XML output filtered by exact clone resource instance =#=#=#=
=#=#=#= End test: XML output filtered by exact clone resource instance - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by exact clone resource instance
=#=#=#= Begin test: Basic text output filtered by resource that doesn't exist =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* No active resources
=#=#=#= End test: Basic text output filtered by resource that doesn't exist - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output filtered by resource that doesn't exist
=#=#=#= Begin test: XML output filtered by resource that doesn't exist =#=#=#=
=#=#=#= End test: XML output filtered by resource that doesn't exist - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by resource that doesn't exist
=#=#=#= Begin test: Basic text output with inactive resources, filtered by tag =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Clone Set: inactive-clone [inactive-dhcpd] (disabled):
* Stopped (disabled): [ cluster01 cluster02 ]
* Resource Group: inactive-group (disabled):
* inactive-dummy-1 (ocf:pacemaker:Dummy): Stopped (disabled)
* inactive-dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
=#=#=#= End test: Basic text output with inactive resources, filtered by tag - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by tag
=#=#=#= Begin test: Basic text output with inactive resources, filtered by bundle resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped
=#=#=#= End test: Basic text output with inactive resources, filtered by bundle resource - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by bundle resource
=#=#=#= Begin test: XML output filtered by inactive bundle resource =#=#=#=
=#=#=#= End test: XML output filtered by inactive bundle resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by inactive bundle resource
=#=#=#= Begin test: Basic text output with inactive resources, filtered by bundled IP address resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Container bundle set: httpd-bundle [pcmk:http]:
* Replica[0]
* httpd-bundle-ip-192.168.122.131 (ocf:heartbeat:IPaddr2): Started cluster01
=#=#=#= End test: Basic text output with inactive resources, filtered by bundled IP address resource - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by bundled IP address resource
=#=#=#= Begin test: XML output filtered by bundled IP address resource =#=#=#=
=#=#=#= End test: XML output filtered by bundled IP address resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by bundled IP address resource
=#=#=#= Begin test: Basic text output with inactive resources, filtered by bundled container =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Container bundle set: httpd-bundle [pcmk:http]:
* Replica[1]
* httpd-bundle-docker-1 (ocf:heartbeat:docker): Started cluster02
=#=#=#= End test: Basic text output with inactive resources, filtered by bundled container - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by bundled container
=#=#=#= Begin test: XML output filtered by bundled container =#=#=#=
=#=#=#= End test: XML output filtered by bundled container - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by bundled container
=#=#=#= Begin test: Basic text output with inactive resources, filtered by bundle connection =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Container bundle set: httpd-bundle [pcmk:http]:
* Replica[0]
* httpd-bundle-0 (ocf:pacemaker:remote): Started cluster01
=#=#=#= End test: Basic text output with inactive resources, filtered by bundle connection - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by bundle connection
=#=#=#= Begin test: XML output filtered by bundle connection =#=#=#=
=#=#=#= End test: XML output filtered by bundle connection - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by bundle connection
=#=#=#= Begin test: Basic text output with inactive resources, filtered by bundled primitive resource =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Full List of Resources:
* Container bundle set: httpd-bundle [pcmk:http]:
* Replica[0]
* httpd (ocf:heartbeat:apache): Started httpd-bundle-0
* Replica[1]
* httpd (ocf:heartbeat:apache): Started httpd-bundle-1
* Replica[2]
* httpd (ocf:heartbeat:apache): Stopped
=#=#=#= End test: Basic text output with inactive resources, filtered by bundled primitive resource - OK (0) =#=#=#=
* Passed: crm_mon - Basic text output with inactive resources, filtered by bundled primitive resource
=#=#=#= Begin test: XML output filtered by bundled primitive resource =#=#=#=
=#=#=#= End test: XML output filtered by bundled primitive resource - OK (0) =#=#=#=
* Passed: crm_mon - XML output filtered by bundled primitive resource
=#=#=#= Begin test: Complete text output, filtered by clone name in cloned group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:0:
* mysql-proxy (lsb:mysql-proxy): Started cluster02
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* Node: cluster01 (1):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
=#=#=#= End test: Complete text output, filtered by clone name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output, filtered by clone name in cloned group
=#=#=#= Begin test: XML output, filtered by clone name in cloned group =#=#=#=
=#=#=#= End test: XML output, filtered by clone name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - XML output, filtered by clone name in cloned group
=#=#=#= Begin test: Complete text output, filtered by group name in cloned group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:0:
* mysql-proxy (lsb:mysql-proxy): Started cluster02
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* Node: cluster01 (1):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
=#=#=#= End test: Complete text output, filtered by group name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output, filtered by group name in cloned group
=#=#=#= Begin test: XML output, filtered by group name in cloned group =#=#=#=
=#=#=#= End test: XML output, filtered by group name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - XML output, filtered by group name in cloned group
=#=#=#= Begin test: Complete text output, filtered by exact group instance name in cloned group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* Node: cluster01 (1):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
=#=#=#= End test: Complete text output, filtered by exact group instance name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output, filtered by exact group instance name in cloned group
=#=#=#= Begin test: XML output, filtered by exact group instance name in cloned group =#=#=#=
=#=#=#= End test: XML output, filtered by exact group instance name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - XML output, filtered by exact group instance name in cloned group
=#=#=#= Begin test: Complete text output, filtered by primitive name in cloned group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:0:
* mysql-proxy (lsb:mysql-proxy): Started cluster02
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* Node: cluster01 (1):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
=#=#=#= End test: Complete text output, filtered by primitive name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output, filtered by primitive name in cloned group
=#=#=#= Begin test: XML output, filtered by primitive name in cloned group =#=#=#=
=#=#=#= End test: XML output, filtered by primitive name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - XML output, filtered by primitive name in cloned group
=#=#=#= Begin test: Complete text output, filtered by exact primitive instance name in cloned group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (2) (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
Node List:
* Online: [ cluster01 (1) cluster02 (2) ]
* GuestOnline: [ httpd-bundle-0@cluster01 httpd-bundle-1@cluster02 ]
Active Resources:
* Clone Set: mysql-clone-group [mysql-group]:
* Resource Group: mysql-group:1:
* mysql-proxy (lsb:mysql-proxy): Started cluster01
Node Attributes:
* Node: cluster01 (1):
* location : office
* pingd : 1000
* Node: cluster02 (2):
* pingd : 1000
Operations:
* Node: cluster02 (2):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* Node: cluster01 (1):
* mysql-proxy: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
=#=#=#= End test: Complete text output, filtered by exact primitive instance name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - Complete text output, filtered by exact primitive instance name in cloned group
=#=#=#= Begin test: XML output, filtered by exact primitive instance name in cloned group =#=#=#=
=#=#=#= End test: XML output, filtered by exact primitive instance name in cloned group - OK (0) =#=#=#=
* Passed: crm_mon - XML output, filtered by exact primitive instance name in cloned group
=#=#=#= Begin test: Text output of partially active resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Active Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Fencing (stonith:fence_xvm): Started cluster01
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Stopped cluster01
* Resource Group: partially-active-group (1 member inactive):
* dummy-1 (ocf:pacemaker:Dummy): Started cluster02
=#=#=#= End test: Text output of partially active resources - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active resources
=#=#=#= Begin test: XML output of partially active resources =#=#=#=
=#=#=#= End test: XML output of partially active resources - OK (0) =#=#=#=
* Passed: crm_mon - XML output of partially active resources
=#=#=#= Begin test: Text output of partially active resources, with inactive resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Full List of Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Stopped: [ cluster02 ]
* Fencing (stonith:fence_xvm): Started cluster01
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Stopped cluster01
* Resource Group: partially-active-group:
* dummy-1 (ocf:pacemaker:Dummy): Started cluster02
* dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
=#=#=#= End test: Text output of partially active resources, with inactive resources - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active resources, with inactive resources
=#=#=#= Begin test: Complete brief text output, with inactive resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Full List of Resources:
* 1/1 (stonith:fence_xvm): Active cluster01
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Stopped: [ cluster02 ]
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Stopped cluster01
* Resource Group: partially-active-group:
* 1/2 (ocf:pacemaker:Dummy): Active cluster02
Node Attributes:
* Node: cluster01:
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* dummy-1: migration-threshold=1000000:
* (2) start
* Node: cluster01:
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* ping: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
=#=#=#= End test: Complete brief text output, with inactive resources - OK (0) =#=#=#=
* Passed: crm_mon - Complete brief text output, with inactive resources
=#=#=#= Begin test: Text output of partially active group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Active Resources:
* Resource Group: partially-active-group (1 member inactive):
* dummy-1 (ocf:pacemaker:Dummy): Started cluster02
=#=#=#= End test: Text output of partially active group - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active group
=#=#=#= Begin test: Text output of partially active group, with inactive resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Full List of Resources:
* Resource Group: partially-active-group:
* dummy-1 (ocf:pacemaker:Dummy): Started cluster02
* dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
=#=#=#= End test: Text output of partially active group, with inactive resources - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active group, with inactive resources
=#=#=#= Begin test: Text output of active member of partially active group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Active Resources:
* Resource Group: partially-active-group (1 member inactive):
* dummy-1 (ocf:pacemaker:Dummy): Started cluster02
=#=#=#= End test: Text output of active member of partially active group - OK (0) =#=#=#=
* Passed: crm_mon - Text output of active member of partially active group
=#=#=#= Begin test: Text output of inactive member of partially active group =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 cluster02 ]
* GuestOnline: [ httpd-bundle-0@cluster02 httpd-bundle-1@cluster01 ]
Active Resources:
* Resource Group: partially-active-group (1 member inactive):
* dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled)
=#=#=#= End test: Text output of inactive member of partially active group - OK (0) =#=#=#=
* Passed: crm_mon - Text output of inactive member of partially active group
=#=#=#= Begin test: Complete brief text output grouped by node, with inactive resources =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Node cluster01: online:
* Resources:
* 1 (ocf:heartbeat:IPaddr2): Active
* 1 (ocf:heartbeat:docker): Active
* 1 (ocf:pacemaker:ping): Active
* 1 (ocf:pacemaker:remote): Active
* 1 (stonith:fence_xvm): Active
* Node cluster02: online:
* Resources:
* 1 (ocf:heartbeat:IPaddr2): Active
* 1 (ocf:heartbeat:docker): Active
* 1 (ocf:pacemaker:Dummy): Active
* 1 (ocf:pacemaker:remote): Active
* GuestNode httpd-bundle-0@cluster02: online:
* Resources:
* 1 (ocf:heartbeat:apache): Active
Inactive Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Stopped: [ cluster02 ]
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster02
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Stopped cluster01
* Resource Group: partially-active-group:
* 1/2 (ocf:pacemaker:Dummy): Active cluster02
Node Attributes:
* Node: cluster01:
* pingd : 1000
* Node: cluster02:
* pingd : 1000
Operations:
* Node: cluster02:
* httpd-bundle-ip-192.168.122.131: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-0: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* dummy-1: migration-threshold=1000000:
* (2) start
* Node: cluster01:
* Fencing: migration-threshold=1000000:
* (15) start
* (20) monitor: interval="60000ms"
* ping: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="10000ms"
* httpd-bundle-ip-192.168.122.132: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-docker-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="60000ms"
* httpd-bundle-1: migration-threshold=1000000:
* (2) start
* (3) monitor: interval="30000ms"
* Node: httpd-bundle-0@cluster02:
* httpd: migration-threshold=1000000:
* (1) start
=#=#=#= End test: Complete brief text output grouped by node, with inactive resources - OK (0) =#=#=#=
* Passed: crm_mon - Complete brief text output grouped by node, with inactive resources
=#=#=#= Begin test: Text output of partially active resources, with inactive resources, filtered by node =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 4 nodes configured
* 13 resource instances configured (1 DISABLED)
Node List:
* Online: [ cluster01 ]
Full List of Resources:
* Clone Set: ping-clone [ping]:
* Started: [ cluster01 ]
* Fencing (stonith:fence_xvm): Started cluster01
* Container bundle set: httpd-bundle [pcmk:http]:
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Stopped cluster01
=#=#=#= End test: Text output of partially active resources, with inactive resources, filtered by node - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active resources, with inactive resources, filtered by node
=#=#=#= Begin test: Text output of partially active resources, filtered by node =#=#=#=
=#=#=#= End test: Text output of partially active resources, filtered by node - OK (0) =#=#=#=
* Passed: crm_mon - Text output of partially active resources, filtered by node
=#=#=#= Begin test: Text output of all resources with maintenance-mode enabled =#=#=#=
Cluster Summary:
* Stack: corosync
* Current DC: cluster02 (version) - partition with quorum
* Last updated:
* Last change:
* 5 nodes configured
* 32 resource instances configured (4 DISABLED)
*** Resource management is DISABLED ***
The cluster will not attempt to start, stop or recover services
Node List:
* GuestNode httpd-bundle-0@cluster01: maintenance
* GuestNode httpd-bundle-1@cluster02: maintenance
* Online: [ cluster01 cluster02 ]
Full List of Resources:
* Clone Set: ping-clone [ping] (unmanaged):
* ping (ocf:pacemaker:ping): Started cluster02 (unmanaged)
* ping (ocf:pacemaker:ping): Started cluster01 (unmanaged)
* Fencing (stonith:fence_xvm): Started cluster01 (unmanaged)
* dummy (ocf:pacemaker:Dummy): Started cluster02 (unmanaged)
- * Clone Set: inactive-clone [inactive-dhcpd] (unmanaged) (disabled):
+ * Clone Set: inactive-clone [inactive-dhcpd] (unmanaged, disabled):
* Stopped (disabled): [ cluster01 cluster02 ]
- * Resource Group: inactive-group (unmanaged) (disabled):
+ * Resource Group: inactive-group (unmanaged, disabled):
* inactive-dummy-1 (ocf:pacemaker:Dummy): Stopped (disabled, unmanaged)
* inactive-dummy-2 (ocf:pacemaker:Dummy): Stopped (disabled, unmanaged)
* Container bundle set: httpd-bundle [pcmk:http] (unmanaged):
* httpd-bundle-0 (192.168.122.131) (ocf:heartbeat:apache): Started cluster01 (unmanaged)
* httpd-bundle-1 (192.168.122.132) (ocf:heartbeat:apache): Started cluster02 (unmanaged)
* httpd-bundle-2 (192.168.122.133) (ocf:heartbeat:apache): Stopped (unmanaged)
* Resource Group: exim-group (unmanaged):
* Public-IP (ocf:heartbeat:IPaddr): Started cluster02 (unmanaged)
* Email (lsb:exim): Started cluster02 (unmanaged)
* Clone Set: mysql-clone-group [mysql-group] (unmanaged):
* Resource Group: mysql-group:0 (unmanaged):
* mysql-proxy (lsb:mysql-proxy): Started cluster02 (unmanaged)
* Resource Group: mysql-group:1 (unmanaged):
* mysql-proxy (lsb:mysql-proxy): Started cluster01 (unmanaged)
- * Clone Set: promotable-clone [promotable-rsc] (promotable) (unmanaged):
+ * Clone Set: promotable-clone [promotable-rsc] (promotable, unmanaged):
* promotable-rsc (ocf:pacemaker:Stateful): Promoted cluster02 (unmanaged)
* promotable-rsc (ocf:pacemaker:Stateful): Unpromoted cluster01 (unmanaged)
=#=#=#= End test: Text output of all resources with maintenance-mode enabled - OK (0) =#=#=#=
* Passed: crm_mon - Text output of all resources with maintenance-mode enabled
diff --git a/cts/scheduler/summary/bug-1822.summary b/cts/scheduler/summary/bug-1822.summary
index f0e4b9b0d1..00457a9c39 100644
--- a/cts/scheduler/summary/bug-1822.summary
+++ b/cts/scheduler/summary/bug-1822.summary
@@ -1,44 +1,44 @@
Current cluster status:
* Node List:
* Online: [ process1a process2b ]
* Full List of Resources:
- * Clone Set: ms-sf [ms-sf_group] (promotable) (unique):
+ * Clone Set: ms-sf [ms-sf_group] (promotable, unique):
* Resource Group: ms-sf_group:0:
* master_slave_Stateful:0 (ocf:heartbeat:Dummy-statful): Unpromoted process2b
* master_slave_procdctl:0 (ocf:heartbeat:procdctl): Stopped
* Resource Group: ms-sf_group:1:
* master_slave_Stateful:1 (ocf:heartbeat:Dummy-statful): Promoted process1a
* master_slave_procdctl:1 (ocf:heartbeat:procdctl): Promoted process1a
Transition Summary:
* Stop master_slave_Stateful:1 ( Promoted process1a ) due to node availability
* Stop master_slave_procdctl:1 ( Promoted process1a ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms-sf_demote_0
* Pseudo action: ms-sf_group:1_demote_0
* Resource action: master_slave_Stateful:1 demote on process1a
* Resource action: master_slave_procdctl:1 demote on process1a
* Pseudo action: ms-sf_group:1_demoted_0
* Pseudo action: ms-sf_demoted_0
* Pseudo action: ms-sf_stop_0
* Pseudo action: ms-sf_group:1_stop_0
* Resource action: master_slave_Stateful:1 stop on process1a
* Resource action: master_slave_procdctl:1 stop on process1a
* Cluster action: do_shutdown on process1a
* Pseudo action: ms-sf_group:1_stopped_0
* Pseudo action: ms-sf_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ process1a process2b ]
* Full List of Resources:
- * Clone Set: ms-sf [ms-sf_group] (promotable) (unique):
+ * Clone Set: ms-sf [ms-sf_group] (promotable, unique):
* Resource Group: ms-sf_group:0:
* master_slave_Stateful:0 (ocf:heartbeat:Dummy-statful): Unpromoted process2b
* master_slave_procdctl:0 (ocf:heartbeat:procdctl): Stopped
* Resource Group: ms-sf_group:1:
* master_slave_Stateful:1 (ocf:heartbeat:Dummy-statful): Stopped
* master_slave_procdctl:1 (ocf:heartbeat:procdctl): Stopped
diff --git a/cts/scheduler/summary/bug-5140-require-all-false.summary b/cts/scheduler/summary/bug-5140-require-all-false.summary
index 9623dd82da..a56fe6d6cc 100644
--- a/cts/scheduler/summary/bug-5140-require-all-false.summary
+++ b/cts/scheduler/summary/bug-5140-require-all-false.summary
@@ -1,83 +1,83 @@
4 of 35 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Node hex-1: standby
* Node hex-2: standby (with active resources)
* Node hex-3: OFFLINE (standby)
* Full List of Resources:
* fencing (stonith:external/sbd): Stopped
* Clone Set: baseclone [basegrp]:
* Resource Group: basegrp:0:
* dlm (ocf:pacemaker:controld): Started hex-2
* clvmd (ocf:lvm2:clvmd): Started hex-2
* o2cb (ocf:ocfs2:o2cb): Started hex-2
* vg1 (ocf:heartbeat:LVM): Stopped
* fs-ocfs-1 (ocf:heartbeat:Filesystem): Stopped
* Stopped: [ hex-1 hex-3 ]
* fs-xfs-1 (ocf:heartbeat:Filesystem): Stopped
* Clone Set: fs2 [fs-ocfs-2]:
* Stopped: [ hex-1 hex-2 hex-3 ]
- * Clone Set: ms-r0 [drbd-r0] (promotable) (disabled):
+ * Clone Set: ms-r0 [drbd-r0] (promotable, disabled):
* Stopped (disabled): [ hex-1 hex-2 hex-3 ]
- * Clone Set: ms-r1 [drbd-r1] (promotable) (disabled):
+ * Clone Set: ms-r1 [drbd-r1] (promotable, disabled):
* Stopped (disabled): [ hex-1 hex-2 hex-3 ]
* Resource Group: md0-group:
* md0 (ocf:heartbeat:Raid1): Stopped
* vg-md0 (ocf:heartbeat:LVM): Stopped
* fs-md0 (ocf:heartbeat:Filesystem): Stopped
* dummy1 (ocf:heartbeat:Delay): Stopped
* dummy3 (ocf:heartbeat:Delay): Stopped
* dummy4 (ocf:heartbeat:Delay): Stopped
* dummy5 (ocf:heartbeat:Delay): Stopped
* dummy6 (ocf:heartbeat:Delay): Stopped
* Resource Group: r0-group:
* fs-r0 (ocf:heartbeat:Filesystem): Stopped
* dummy2 (ocf:heartbeat:Delay): Stopped
* cluster-md0 (ocf:heartbeat:Raid1): Stopped
Transition Summary:
* Stop dlm:0 ( hex-2 ) due to node availability
* Stop clvmd:0 ( hex-2 ) due to node availability
* Stop o2cb:0 ( hex-2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: baseclone_stop_0
* Pseudo action: basegrp:0_stop_0
* Resource action: o2cb stop on hex-2
* Resource action: clvmd stop on hex-2
* Resource action: dlm stop on hex-2
* Pseudo action: basegrp:0_stopped_0
* Pseudo action: baseclone_stopped_0
Revised Cluster Status:
* Node List:
* Node hex-1: standby
* Node hex-2: standby
* Node hex-3: OFFLINE (standby)
* Full List of Resources:
* fencing (stonith:external/sbd): Stopped
* Clone Set: baseclone [basegrp]:
* Stopped: [ hex-1 hex-2 hex-3 ]
* fs-xfs-1 (ocf:heartbeat:Filesystem): Stopped
* Clone Set: fs2 [fs-ocfs-2]:
* Stopped: [ hex-1 hex-2 hex-3 ]
- * Clone Set: ms-r0 [drbd-r0] (promotable) (disabled):
+ * Clone Set: ms-r0 [drbd-r0] (promotable, disabled):
* Stopped (disabled): [ hex-1 hex-2 hex-3 ]
- * Clone Set: ms-r1 [drbd-r1] (promotable) (disabled):
+ * Clone Set: ms-r1 [drbd-r1] (promotable, disabled):
* Stopped (disabled): [ hex-1 hex-2 hex-3 ]
* Resource Group: md0-group:
* md0 (ocf:heartbeat:Raid1): Stopped
* vg-md0 (ocf:heartbeat:LVM): Stopped
* fs-md0 (ocf:heartbeat:Filesystem): Stopped
* dummy1 (ocf:heartbeat:Delay): Stopped
* dummy3 (ocf:heartbeat:Delay): Stopped
* dummy4 (ocf:heartbeat:Delay): Stopped
* dummy5 (ocf:heartbeat:Delay): Stopped
* dummy6 (ocf:heartbeat:Delay): Stopped
* Resource Group: r0-group:
* fs-r0 (ocf:heartbeat:Filesystem): Stopped
* dummy2 (ocf:heartbeat:Delay): Stopped
* cluster-md0 (ocf:heartbeat:Raid1): Stopped
diff --git a/cts/scheduler/summary/bug-lf-2358.summary b/cts/scheduler/summary/bug-lf-2358.summary
index 7c2c3d220b..b89aadc317 100644
--- a/cts/scheduler/summary/bug-lf-2358.summary
+++ b/cts/scheduler/summary/bug-lf-2358.summary
@@ -1,68 +1,68 @@
2 of 15 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Online: [ alice.demo bob.demo ]
* Full List of Resources:
- * Clone Set: ms_drbd_nfsexport [res_drbd_nfsexport] (promotable) (disabled):
+ * Clone Set: ms_drbd_nfsexport [res_drbd_nfsexport] (promotable, disabled):
* Stopped (disabled): [ alice.demo bob.demo ]
* Resource Group: rg_nfs:
* res_fs_nfsexport (ocf:heartbeat:Filesystem): Stopped
* res_ip_nfs (ocf:heartbeat:IPaddr2): Stopped
* res_nfs (lsb:nfs): Stopped
* Resource Group: rg_mysql1:
* res_fs_mysql1 (ocf:heartbeat:Filesystem): Started bob.demo
* res_ip_mysql1 (ocf:heartbeat:IPaddr2): Started bob.demo
* res_mysql1 (ocf:heartbeat:mysql): Started bob.demo
* Clone Set: ms_drbd_mysql1 [res_drbd_mysql1] (promotable):
* Promoted: [ bob.demo ]
* Stopped: [ alice.demo ]
* Clone Set: ms_drbd_mysql2 [res_drbd_mysql2] (promotable):
* Promoted: [ alice.demo ]
* Unpromoted: [ bob.demo ]
* Resource Group: rg_mysql2:
* res_fs_mysql2 (ocf:heartbeat:Filesystem): Started alice.demo
* res_ip_mysql2 (ocf:heartbeat:IPaddr2): Started alice.demo
* res_mysql2 (ocf:heartbeat:mysql): Started alice.demo
Transition Summary:
* Start res_drbd_mysql1:1 ( alice.demo )
Executing Cluster Transition:
* Pseudo action: ms_drbd_mysql1_pre_notify_start_0
* Resource action: res_drbd_mysql1:0 notify on bob.demo
* Pseudo action: ms_drbd_mysql1_confirmed-pre_notify_start_0
* Pseudo action: ms_drbd_mysql1_start_0
* Resource action: res_drbd_mysql1:1 start on alice.demo
* Pseudo action: ms_drbd_mysql1_running_0
* Pseudo action: ms_drbd_mysql1_post_notify_running_0
* Resource action: res_drbd_mysql1:0 notify on bob.demo
* Resource action: res_drbd_mysql1:1 notify on alice.demo
* Pseudo action: ms_drbd_mysql1_confirmed-post_notify_running_0
Revised Cluster Status:
* Node List:
* Online: [ alice.demo bob.demo ]
* Full List of Resources:
- * Clone Set: ms_drbd_nfsexport [res_drbd_nfsexport] (promotable) (disabled):
+ * Clone Set: ms_drbd_nfsexport [res_drbd_nfsexport] (promotable, disabled):
* Stopped (disabled): [ alice.demo bob.demo ]
* Resource Group: rg_nfs:
* res_fs_nfsexport (ocf:heartbeat:Filesystem): Stopped
* res_ip_nfs (ocf:heartbeat:IPaddr2): Stopped
* res_nfs (lsb:nfs): Stopped
* Resource Group: rg_mysql1:
* res_fs_mysql1 (ocf:heartbeat:Filesystem): Started bob.demo
* res_ip_mysql1 (ocf:heartbeat:IPaddr2): Started bob.demo
* res_mysql1 (ocf:heartbeat:mysql): Started bob.demo
* Clone Set: ms_drbd_mysql1 [res_drbd_mysql1] (promotable):
* Promoted: [ bob.demo ]
* Unpromoted: [ alice.demo ]
* Clone Set: ms_drbd_mysql2 [res_drbd_mysql2] (promotable):
* Promoted: [ alice.demo ]
* Unpromoted: [ bob.demo ]
* Resource Group: rg_mysql2:
* res_fs_mysql2 (ocf:heartbeat:Filesystem): Started alice.demo
* res_ip_mysql2 (ocf:heartbeat:IPaddr2): Started alice.demo
* res_mysql2 (ocf:heartbeat:mysql): Started alice.demo
diff --git a/cts/scheduler/summary/bug-pm-11.summary b/cts/scheduler/summary/bug-pm-11.summary
index f638b3fc4f..7a9fc5c1b0 100644
--- a/cts/scheduler/summary/bug-pm-11.summary
+++ b/cts/scheduler/summary/bug-pm-11.summary
@@ -1,48 +1,48 @@
Current cluster status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Stopped
Transition Summary:
* Start stateful-2:0 ( node-b )
* Promote stateful-2:1 ( Stopped -> Promoted node-a )
Executing Cluster Transition:
* Resource action: stateful-2:0 monitor on node-b
* Resource action: stateful-2:0 monitor on node-a
* Resource action: stateful-2:1 monitor on node-b
* Resource action: stateful-2:1 monitor on node-a
* Pseudo action: ms-sf_start_0
* Pseudo action: group:0_start_0
* Resource action: stateful-2:0 start on node-b
* Pseudo action: group:1_start_0
* Resource action: stateful-2:1 start on node-a
* Pseudo action: group:0_running_0
* Pseudo action: group:1_running_0
* Pseudo action: ms-sf_running_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-2:1 promote on node-a
* Pseudo action: group:1_promoted_0
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a
diff --git a/cts/scheduler/summary/bug-pm-12.summary b/cts/scheduler/summary/bug-pm-12.summary
index c4f3adb908..2b473e8b91 100644
--- a/cts/scheduler/summary/bug-pm-12.summary
+++ b/cts/scheduler/summary/bug-pm-12.summary
@@ -1,57 +1,57 @@
Current cluster status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a
Transition Summary:
* Restart stateful-2:0 ( Unpromoted node-b ) due to resource definition change
* Restart stateful-2:1 ( Promoted node-a ) due to resource definition change
Executing Cluster Transition:
* Pseudo action: ms-sf_demote_0
* Pseudo action: group:1_demote_0
* Resource action: stateful-2:1 demote on node-a
* Pseudo action: group:1_demoted_0
* Pseudo action: ms-sf_demoted_0
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-2:0 stop on node-b
* Pseudo action: group:1_stop_0
* Resource action: stateful-2:1 stop on node-a
* Pseudo action: group:0_stopped_0
* Pseudo action: group:1_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_start_0
* Pseudo action: group:0_start_0
* Resource action: stateful-2:0 start on node-b
* Pseudo action: group:1_start_0
* Resource action: stateful-2:1 start on node-a
* Pseudo action: group:0_running_0
* Pseudo action: group:1_running_0
* Pseudo action: ms-sf_running_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-2:1 promote on node-a
* Pseudo action: group:1_promoted_0
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a
diff --git a/cts/scheduler/summary/group14.summary b/cts/scheduler/summary/group14.summary
index a1ba66a2e5..80ded38d78 100644
--- a/cts/scheduler/summary/group14.summary
+++ b/cts/scheduler/summary/group14.summary
@@ -1,102 +1,102 @@
Current cluster status:
* Node List:
* Online: [ c001n06 c001n07 ]
* OFFLINE: [ c001n02 c001n03 c001n04 c001n05 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n06
* r192.168.100.182 (ocf:heartbeat:IPaddr): Stopped
* r192.168.100.183 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Stopped
* migrator (ocf:heartbeat:Dummy): Stopped
* rsc_c001n03 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n02 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n04 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n05 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n06 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n07 (ocf:heartbeat:IPaddr): Stopped
* Clone Set: DoFencing [child_DoFencing]:
* Stopped: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Stopped
Transition Summary:
* Start DcIPaddr ( c001n06 ) due to no quorum (blocked)
* Stop r192.168.100.181 ( c001n06 ) due to no quorum
* Start r192.168.100.182 ( c001n07 ) due to no quorum (blocked)
* Start r192.168.100.183 ( c001n07 ) due to no quorum (blocked)
* Start lsb_dummy ( c001n06 ) due to no quorum (blocked)
* Start migrator ( c001n06 ) due to no quorum (blocked)
* Start rsc_c001n03 ( c001n06 ) due to no quorum (blocked)
* Start rsc_c001n02 ( c001n07 ) due to no quorum (blocked)
* Start rsc_c001n04 ( c001n06 ) due to no quorum (blocked)
* Start rsc_c001n05 ( c001n07 ) due to no quorum (blocked)
* Start rsc_c001n06 ( c001n06 ) due to no quorum (blocked)
* Start rsc_c001n07 ( c001n07 ) due to no quorum (blocked)
* Start child_DoFencing:0 ( c001n06 )
* Start child_DoFencing:1 ( c001n07 )
* Start ocf_msdummy:0 ( c001n06 ) due to no quorum (blocked)
* Start ocf_msdummy:1 ( c001n07 ) due to no quorum (blocked)
* Start ocf_msdummy:2 ( c001n06 ) due to no quorum (blocked)
* Start ocf_msdummy:3 ( c001n07 ) due to no quorum (blocked)
Executing Cluster Transition:
* Pseudo action: group-1_stop_0
* Resource action: r192.168.100.181 stop on c001n06
* Pseudo action: DoFencing_start_0
* Pseudo action: group-1_stopped_0
* Pseudo action: group-1_start_0
* Resource action: child_DoFencing:0 start on c001n06
* Resource action: child_DoFencing:1 start on c001n07
* Pseudo action: DoFencing_running_0
* Resource action: child_DoFencing:0 monitor=20000 on c001n06
* Resource action: child_DoFencing:1 monitor=20000 on c001n07
Revised Cluster Status:
* Node List:
* Online: [ c001n06 c001n07 ]
* OFFLINE: [ c001n02 c001n03 c001n04 c001n05 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Stopped
* r192.168.100.182 (ocf:heartbeat:IPaddr): Stopped
* r192.168.100.183 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Stopped
* migrator (ocf:heartbeat:Dummy): Stopped
* rsc_c001n03 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n02 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n04 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n05 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n06 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n07 (ocf:heartbeat:IPaddr): Stopped
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n06 c001n07 ]
* Stopped: [ c001n02 c001n03 c001n04 c001n05 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Stopped
diff --git a/cts/scheduler/summary/inc11.summary b/cts/scheduler/summary/inc11.summary
index 51e838c374..256a10e8f7 100644
--- a/cts/scheduler/summary/inc11.summary
+++ b/cts/scheduler/summary/inc11.summary
@@ -1,43 +1,43 @@
Current cluster status:
* Node List:
* Online: [ node0 node1 node2 ]
* Full List of Resources:
* simple-rsc (ocf:heartbeat:apache): Stopped
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Start simple-rsc ( node2 )
* Start child_rsc1:0 ( node1 )
* Promote child_rsc1:1 ( Stopped -> Promoted node2 )
Executing Cluster Transition:
* Resource action: simple-rsc monitor on node2
* Resource action: simple-rsc monitor on node1
* Resource action: simple-rsc monitor on node0
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:0 monitor on node0
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:1 monitor on node0
* Pseudo action: rsc1_start_0
* Resource action: simple-rsc start on node2
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:1 promote on node2
* Pseudo action: rsc1_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node0 node1 node2 ]
* Full List of Resources:
* simple-rsc (ocf:heartbeat:apache): Started node2
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Promoted node2
diff --git a/cts/scheduler/summary/inc12.summary b/cts/scheduler/summary/inc12.summary
index 1ada08dda0..2c93e2678c 100644
--- a/cts/scheduler/summary/inc12.summary
+++ b/cts/scheduler/summary/inc12.summary
@@ -1,132 +1,132 @@
Current cluster status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n04
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n02 c001n04 c001n05 c001n06 c001n07 ]
* Stopped: [ c001n03 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Unpromoted c001n04
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Unpromoted c001n04
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Unpromoted c001n05
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Unpromoted c001n05
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Unpromoted c001n06
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Unpromoted c001n06
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Unpromoted c001n07
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Unpromoted c001n07
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Unpromoted c001n02
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Unpromoted c001n02
Transition Summary:
* Stop ocf_192.168.100.181 ( c001n02 ) due to node availability
* Stop heartbeat_192.168.100.182 ( c001n02 ) due to node availability
* Stop ocf_192.168.100.183 ( c001n02 ) due to node availability
* Stop lsb_dummy ( c001n04 ) due to node availability
* Stop rsc_c001n03 ( c001n05 ) due to node availability
* Stop rsc_c001n02 ( c001n02 ) due to node availability
* Stop rsc_c001n04 ( c001n04 ) due to node availability
* Stop rsc_c001n05 ( c001n05 ) due to node availability
* Stop rsc_c001n06 ( c001n06 ) due to node availability
* Stop rsc_c001n07 ( c001n07 ) due to node availability
* Stop child_DoFencing:0 ( c001n02 ) due to node availability
* Stop child_DoFencing:1 ( c001n04 ) due to node availability
* Stop child_DoFencing:2 ( c001n05 ) due to node availability
* Stop child_DoFencing:3 ( c001n06 ) due to node availability
* Stop child_DoFencing:4 ( c001n07 ) due to node availability
* Stop ocf_msdummy:2 ( Unpromoted c001n04 ) due to node availability
* Stop ocf_msdummy:3 ( Unpromoted c001n04 ) due to node availability
* Stop ocf_msdummy:4 ( Unpromoted c001n05 ) due to node availability
* Stop ocf_msdummy:5 ( Unpromoted c001n05 ) due to node availability
* Stop ocf_msdummy:6 ( Unpromoted c001n06 ) due to node availability
* Stop ocf_msdummy:7 ( Unpromoted c001n06 ) due to node availability
* Stop ocf_msdummy:8 ( Unpromoted c001n07 ) due to node availability
* Stop ocf_msdummy:9 ( Unpromoted c001n07 ) due to node availability
* Stop ocf_msdummy:10 ( Unpromoted c001n02 ) due to node availability
* Stop ocf_msdummy:11 ( Unpromoted c001n02 ) due to node availability
Executing Cluster Transition:
* Pseudo action: group-1_stop_0
* Resource action: ocf_192.168.100.183 stop on c001n02
* Resource action: lsb_dummy stop on c001n04
* Resource action: rsc_c001n03 stop on c001n05
* Resource action: rsc_c001n02 stop on c001n02
* Resource action: rsc_c001n04 stop on c001n04
* Resource action: rsc_c001n05 stop on c001n05
* Resource action: rsc_c001n06 stop on c001n06
* Resource action: rsc_c001n07 stop on c001n07
* Pseudo action: DoFencing_stop_0
* Pseudo action: master_rsc_1_stop_0
* Resource action: heartbeat_192.168.100.182 stop on c001n02
* Resource action: child_DoFencing:1 stop on c001n02
* Resource action: child_DoFencing:2 stop on c001n04
* Resource action: child_DoFencing:3 stop on c001n05
* Resource action: child_DoFencing:4 stop on c001n06
* Resource action: child_DoFencing:5 stop on c001n07
* Pseudo action: DoFencing_stopped_0
* Resource action: ocf_msdummy:2 stop on c001n04
* Resource action: ocf_msdummy:3 stop on c001n04
* Resource action: ocf_msdummy:4 stop on c001n05
* Resource action: ocf_msdummy:5 stop on c001n05
* Resource action: ocf_msdummy:6 stop on c001n06
* Resource action: ocf_msdummy:7 stop on c001n06
* Resource action: ocf_msdummy:8 stop on c001n07
* Resource action: ocf_msdummy:9 stop on c001n07
* Resource action: ocf_msdummy:10 stop on c001n02
* Resource action: ocf_msdummy:11 stop on c001n02
* Pseudo action: master_rsc_1_stopped_0
* Cluster action: do_shutdown on c001n07
* Cluster action: do_shutdown on c001n06
* Cluster action: do_shutdown on c001n05
* Cluster action: do_shutdown on c001n04
* Resource action: ocf_192.168.100.181 stop on c001n02
* Cluster action: do_shutdown on c001n02
* Pseudo action: group-1_stopped_0
* Cluster action: do_shutdown on c001n03
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Stopped
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Stopped
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Stopped
* rsc_c001n03 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n02 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n04 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n05 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n06 (ocf:heartbeat:IPaddr): Stopped
* rsc_c001n07 (ocf:heartbeat:IPaddr): Stopped
* Clone Set: DoFencing [child_DoFencing]:
* Stopped: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Stopped
diff --git a/cts/scheduler/summary/managed-1.summary b/cts/scheduler/summary/managed-1.summary
index 1c417a4fb5..9c25080237 100644
--- a/cts/scheduler/summary/managed-1.summary
+++ b/cts/scheduler/summary/managed-1.summary
@@ -1,132 +1,132 @@
Current cluster status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 c001n08 c001n09 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n09 (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
- * Clone Set: DoFencing [child_DoFencing] (unique) (unmanaged):
+ * Clone Set: DoFencing [child_DoFencing] (unique, unmanaged):
* child_DoFencing:0 (stonith:ssh): Started c001n02 (unmanaged)
* child_DoFencing:1 (stonith:ssh): Started c001n03 (unmanaged)
* child_DoFencing:2 (stonith:ssh): Started c001n04 (unmanaged)
* child_DoFencing:3 (stonith:ssh): Started c001n05 (unmanaged)
* child_DoFencing:4 (stonith:ssh): Started c001n06 (unmanaged)
* child_DoFencing:5 (stonith:ssh): Started c001n07 (unmanaged)
* child_DoFencing:6 (stonith:ssh): Started c001n08 (unmanaged)
* child_DoFencing:7 (stonith:ssh): Started c001n09 (unmanaged)
Transition Summary:
Executing Cluster Transition:
* Resource action: DcIPaddr monitor on c001n08
* Resource action: DcIPaddr monitor on c001n07
* Resource action: DcIPaddr monitor on c001n06
* Resource action: DcIPaddr monitor on c001n05
* Resource action: DcIPaddr monitor on c001n04
* Resource action: DcIPaddr monitor on c001n03
* Resource action: DcIPaddr monitor on c001n02
* Resource action: rsc_c001n09 monitor on c001n08
* Resource action: rsc_c001n09 monitor on c001n07
* Resource action: rsc_c001n09 monitor on c001n05
* Resource action: rsc_c001n09 monitor on c001n04
* Resource action: rsc_c001n09 monitor on c001n03
* Resource action: rsc_c001n09 monitor on c001n02
* Resource action: rsc_c001n02 monitor on c001n09
* Resource action: rsc_c001n02 monitor on c001n08
* Resource action: rsc_c001n02 monitor on c001n07
* Resource action: rsc_c001n02 monitor on c001n05
* Resource action: rsc_c001n02 monitor on c001n04
* Resource action: rsc_c001n03 monitor on c001n09
* Resource action: rsc_c001n03 monitor on c001n08
* Resource action: rsc_c001n03 monitor on c001n07
* Resource action: rsc_c001n03 monitor on c001n05
* Resource action: rsc_c001n03 monitor on c001n04
* Resource action: rsc_c001n03 monitor on c001n02
* Resource action: rsc_c001n04 monitor on c001n09
* Resource action: rsc_c001n04 monitor on c001n08
* Resource action: rsc_c001n04 monitor on c001n07
* Resource action: rsc_c001n04 monitor on c001n05
* Resource action: rsc_c001n04 monitor on c001n03
* Resource action: rsc_c001n04 monitor on c001n02
* Resource action: rsc_c001n05 monitor on c001n09
* Resource action: rsc_c001n05 monitor on c001n08
* Resource action: rsc_c001n05 monitor on c001n07
* Resource action: rsc_c001n05 monitor on c001n06
* Resource action: rsc_c001n05 monitor on c001n04
* Resource action: rsc_c001n05 monitor on c001n03
* Resource action: rsc_c001n05 monitor on c001n02
* Resource action: rsc_c001n06 monitor on c001n09
* Resource action: rsc_c001n06 monitor on c001n08
* Resource action: rsc_c001n06 monitor on c001n07
* Resource action: rsc_c001n06 monitor on c001n05
* Resource action: rsc_c001n06 monitor on c001n04
* Resource action: rsc_c001n06 monitor on c001n03
* Resource action: rsc_c001n07 monitor on c001n09
* Resource action: rsc_c001n07 monitor on c001n08
* Resource action: rsc_c001n07 monitor on c001n06
* Resource action: rsc_c001n07 monitor on c001n05
* Resource action: rsc_c001n07 monitor on c001n04
* Resource action: rsc_c001n08 monitor on c001n09
* Resource action: rsc_c001n08 monitor on c001n07
* Resource action: rsc_c001n08 monitor on c001n05
* Resource action: child_DoFencing:0 monitor on c001n09
* Resource action: child_DoFencing:0 monitor on c001n08
* Resource action: child_DoFencing:0 monitor on c001n07
* Resource action: child_DoFencing:1 monitor on c001n08
* Resource action: child_DoFencing:1 monitor on c001n07
* Resource action: child_DoFencing:1 monitor on c001n02
* Resource action: child_DoFencing:2 monitor on c001n09
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n07
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n08
* Resource action: child_DoFencing:3 monitor on c001n04
* Resource action: child_DoFencing:3 monitor on c001n02
* Resource action: child_DoFencing:4 monitor on c001n09
* Resource action: child_DoFencing:4 monitor on c001n05
* Resource action: child_DoFencing:4 monitor on c001n03
* Resource action: child_DoFencing:5 monitor on c001n08
* Resource action: child_DoFencing:5 monitor on c001n05
* Resource action: child_DoFencing:5 monitor on c001n04
* Resource action: child_DoFencing:5 monitor on c001n02
* Resource action: child_DoFencing:6 monitor on c001n09
* Resource action: child_DoFencing:6 monitor on c001n07
* Resource action: child_DoFencing:6 monitor on c001n05
* Resource action: child_DoFencing:6 monitor on c001n04
* Resource action: child_DoFencing:7 monitor on c001n08
* Resource action: child_DoFencing:7 monitor on c001n07
* Resource action: child_DoFencing:7 monitor on c001n05
* Resource action: child_DoFencing:7 monitor on c001n04
* Resource action: child_DoFencing:7 monitor on c001n03
* Resource action: child_DoFencing:7 monitor on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 c001n08 c001n09 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n09 (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
- * Clone Set: DoFencing [child_DoFencing] (unique) (unmanaged):
+ * Clone Set: DoFencing [child_DoFencing] (unique, unmanaged):
* child_DoFencing:0 (stonith:ssh): Started c001n02 (unmanaged)
* child_DoFencing:1 (stonith:ssh): Started c001n03 (unmanaged)
* child_DoFencing:2 (stonith:ssh): Started c001n04 (unmanaged)
* child_DoFencing:3 (stonith:ssh): Started c001n05 (unmanaged)
* child_DoFencing:4 (stonith:ssh): Started c001n06 (unmanaged)
* child_DoFencing:5 (stonith:ssh): Started c001n07 (unmanaged)
* child_DoFencing:6 (stonith:ssh): Started c001n08 (unmanaged)
* child_DoFencing:7 (stonith:ssh): Started c001n09 (unmanaged)
diff --git a/cts/scheduler/summary/managed-2.summary b/cts/scheduler/summary/managed-2.summary
index a1d327c3da..dd0a1870b8 100644
--- a/cts/scheduler/summary/managed-2.summary
+++ b/cts/scheduler/summary/managed-2.summary
@@ -1,166 +1,166 @@
Current cluster status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 c001n08 c001n09 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n09 (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
- * Clone Set: DoFencing [child_DoFencing] (unique) (unmanaged):
+ * Clone Set: DoFencing [child_DoFencing] (unique, unmanaged):
* child_DoFencing:0 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:1 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:2 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:3 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:4 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:5 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:6 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:7 (stonith:ssh): Stopped (unmanaged)
Transition Summary:
Executing Cluster Transition:
* Resource action: DcIPaddr monitor on c001n08
* Resource action: DcIPaddr monitor on c001n07
* Resource action: DcIPaddr monitor on c001n06
* Resource action: DcIPaddr monitor on c001n05
* Resource action: DcIPaddr monitor on c001n04
* Resource action: DcIPaddr monitor on c001n03
* Resource action: DcIPaddr monitor on c001n02
* Resource action: rsc_c001n09 monitor on c001n08
* Resource action: rsc_c001n09 monitor on c001n07
* Resource action: rsc_c001n09 monitor on c001n05
* Resource action: rsc_c001n09 monitor on c001n04
* Resource action: rsc_c001n09 monitor on c001n03
* Resource action: rsc_c001n09 monitor on c001n02
* Resource action: rsc_c001n02 monitor on c001n09
* Resource action: rsc_c001n02 monitor on c001n08
* Resource action: rsc_c001n02 monitor on c001n07
* Resource action: rsc_c001n02 monitor on c001n05
* Resource action: rsc_c001n02 monitor on c001n04
* Resource action: rsc_c001n03 monitor on c001n09
* Resource action: rsc_c001n03 monitor on c001n08
* Resource action: rsc_c001n03 monitor on c001n07
* Resource action: rsc_c001n03 monitor on c001n05
* Resource action: rsc_c001n03 monitor on c001n04
* Resource action: rsc_c001n03 monitor on c001n02
* Resource action: rsc_c001n04 monitor on c001n09
* Resource action: rsc_c001n04 monitor on c001n08
* Resource action: rsc_c001n04 monitor on c001n07
* Resource action: rsc_c001n04 monitor on c001n05
* Resource action: rsc_c001n04 monitor on c001n03
* Resource action: rsc_c001n04 monitor on c001n02
* Resource action: rsc_c001n05 monitor on c001n09
* Resource action: rsc_c001n05 monitor on c001n08
* Resource action: rsc_c001n05 monitor on c001n07
* Resource action: rsc_c001n05 monitor on c001n06
* Resource action: rsc_c001n05 monitor on c001n04
* Resource action: rsc_c001n05 monitor on c001n03
* Resource action: rsc_c001n05 monitor on c001n02
* Resource action: rsc_c001n06 monitor on c001n09
* Resource action: rsc_c001n06 monitor on c001n08
* Resource action: rsc_c001n06 monitor on c001n07
* Resource action: rsc_c001n06 monitor on c001n05
* Resource action: rsc_c001n06 monitor on c001n04
* Resource action: rsc_c001n06 monitor on c001n03
* Resource action: rsc_c001n07 monitor on c001n09
* Resource action: rsc_c001n07 monitor on c001n08
* Resource action: rsc_c001n07 monitor on c001n06
* Resource action: rsc_c001n07 monitor on c001n05
* Resource action: rsc_c001n07 monitor on c001n04
* Resource action: rsc_c001n08 monitor on c001n09
* Resource action: rsc_c001n08 monitor on c001n07
* Resource action: rsc_c001n08 monitor on c001n05
* Resource action: child_DoFencing:0 monitor on c001n09
* Resource action: child_DoFencing:0 monitor on c001n08
* Resource action: child_DoFencing:0 monitor on c001n07
* Resource action: child_DoFencing:0 monitor on c001n06
* Resource action: child_DoFencing:0 monitor on c001n05
* Resource action: child_DoFencing:0 monitor on c001n04
* Resource action: child_DoFencing:0 monitor on c001n03
* Resource action: child_DoFencing:0 monitor on c001n02
* Resource action: child_DoFencing:1 monitor on c001n09
* Resource action: child_DoFencing:1 monitor on c001n08
* Resource action: child_DoFencing:1 monitor on c001n07
* Resource action: child_DoFencing:1 monitor on c001n06
* Resource action: child_DoFencing:1 monitor on c001n05
* Resource action: child_DoFencing:1 monitor on c001n04
* Resource action: child_DoFencing:1 monitor on c001n03
* Resource action: child_DoFencing:1 monitor on c001n02
* Resource action: child_DoFencing:2 monitor on c001n09
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n07
* Resource action: child_DoFencing:2 monitor on c001n06
* Resource action: child_DoFencing:2 monitor on c001n05
* Resource action: child_DoFencing:2 monitor on c001n04
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:2 monitor on c001n02
* Resource action: child_DoFencing:3 monitor on c001n09
* Resource action: child_DoFencing:3 monitor on c001n08
* Resource action: child_DoFencing:3 monitor on c001n07
* Resource action: child_DoFencing:3 monitor on c001n06
* Resource action: child_DoFencing:3 monitor on c001n05
* Resource action: child_DoFencing:3 monitor on c001n04
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n02
* Resource action: child_DoFencing:4 monitor on c001n09
* Resource action: child_DoFencing:4 monitor on c001n08
* Resource action: child_DoFencing:4 monitor on c001n07
* Resource action: child_DoFencing:4 monitor on c001n06
* Resource action: child_DoFencing:4 monitor on c001n05
* Resource action: child_DoFencing:4 monitor on c001n04
* Resource action: child_DoFencing:4 monitor on c001n03
* Resource action: child_DoFencing:4 monitor on c001n02
* Resource action: child_DoFencing:5 monitor on c001n09
* Resource action: child_DoFencing:5 monitor on c001n08
* Resource action: child_DoFencing:5 monitor on c001n07
* Resource action: child_DoFencing:5 monitor on c001n06
* Resource action: child_DoFencing:5 monitor on c001n05
* Resource action: child_DoFencing:5 monitor on c001n04
* Resource action: child_DoFencing:5 monitor on c001n03
* Resource action: child_DoFencing:5 monitor on c001n02
* Resource action: child_DoFencing:6 monitor on c001n09
* Resource action: child_DoFencing:6 monitor on c001n08
* Resource action: child_DoFencing:6 monitor on c001n07
* Resource action: child_DoFencing:6 monitor on c001n06
* Resource action: child_DoFencing:6 monitor on c001n05
* Resource action: child_DoFencing:6 monitor on c001n04
* Resource action: child_DoFencing:6 monitor on c001n03
* Resource action: child_DoFencing:6 monitor on c001n02
* Resource action: child_DoFencing:7 monitor on c001n09
* Resource action: child_DoFencing:7 monitor on c001n08
* Resource action: child_DoFencing:7 monitor on c001n07
* Resource action: child_DoFencing:7 monitor on c001n06
* Resource action: child_DoFencing:7 monitor on c001n05
* Resource action: child_DoFencing:7 monitor on c001n04
* Resource action: child_DoFencing:7 monitor on c001n03
* Resource action: child_DoFencing:7 monitor on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n04 c001n05 c001n06 c001n07 c001n08 c001n09 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n09 (ocf:heartbeat:IPaddr): Started c001n09
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
- * Clone Set: DoFencing [child_DoFencing] (unique) (unmanaged):
+ * Clone Set: DoFencing [child_DoFencing] (unique, unmanaged):
* child_DoFencing:0 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:1 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:2 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:3 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:4 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:5 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:6 (stonith:ssh): Stopped (unmanaged)
* child_DoFencing:7 (stonith:ssh): Stopped (unmanaged)
diff --git a/cts/scheduler/summary/promoted-0.summary b/cts/scheduler/summary/promoted-0.summary
index b80bb106cf..3e724ffdc4 100644
--- a/cts/scheduler/summary/promoted-0.summary
+++ b/cts/scheduler/summary/promoted-0.summary
@@ -1,47 +1,47 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
* child_rsc1:2 (ocf:heartbeat:apache): Stopped
* child_rsc1:3 (ocf:heartbeat:apache): Stopped
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Start child_rsc1:0 ( node1 )
* Start child_rsc1:1 ( node2 )
* Start child_rsc1:2 ( node1 )
* Start child_rsc1:3 ( node2 )
Executing Cluster Transition:
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:2 monitor on node2
* Resource action: child_rsc1:2 monitor on node1
* Resource action: child_rsc1:3 monitor on node2
* Resource action: child_rsc1:3 monitor on node1
* Resource action: child_rsc1:4 monitor on node2
* Resource action: child_rsc1:4 monitor on node1
* Pseudo action: rsc1_start_0
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Resource action: child_rsc1:2 start on node1
* Resource action: child_rsc1:3 start on node2
* Pseudo action: rsc1_running_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:2 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:3 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
diff --git a/cts/scheduler/summary/promoted-1.summary b/cts/scheduler/summary/promoted-1.summary
index 161f51834a..08100f3e36 100644
--- a/cts/scheduler/summary/promoted-1.summary
+++ b/cts/scheduler/summary/promoted-1.summary
@@ -1,50 +1,50 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
* child_rsc1:2 (ocf:heartbeat:apache): Stopped
* child_rsc1:3 (ocf:heartbeat:apache): Stopped
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Start child_rsc1:0 ( node1 )
* Promote child_rsc1:1 ( Stopped -> Promoted node2 )
* Start child_rsc1:2 ( node1 )
* Start child_rsc1:3 ( node2 )
Executing Cluster Transition:
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:2 monitor on node2
* Resource action: child_rsc1:2 monitor on node1
* Resource action: child_rsc1:3 monitor on node2
* Resource action: child_rsc1:3 monitor on node1
* Resource action: child_rsc1:4 monitor on node2
* Resource action: child_rsc1:4 monitor on node1
* Pseudo action: rsc1_start_0
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Resource action: child_rsc1:2 start on node1
* Resource action: child_rsc1:3 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:1 promote on node2
* Pseudo action: rsc1_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Promoted node2
* child_rsc1:2 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:3 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
diff --git a/cts/scheduler/summary/promoted-10.summary b/cts/scheduler/summary/promoted-10.summary
index 54dbcd7e69..c35c61c793 100644
--- a/cts/scheduler/summary/promoted-10.summary
+++ b/cts/scheduler/summary/promoted-10.summary
@@ -1,75 +1,75 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
* child_rsc1:2 (ocf:heartbeat:apache): Stopped
* child_rsc1:3 (ocf:heartbeat:apache): Stopped
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Promote child_rsc1:0 ( Stopped -> Promoted node1 )
* Start child_rsc1:1 ( node2 )
* Start child_rsc1:2 ( node1 )
* Promote child_rsc1:3 ( Stopped -> Promoted node2 )
Executing Cluster Transition:
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:2 monitor on node2
* Resource action: child_rsc1:2 monitor on node1
* Resource action: child_rsc1:3 monitor on node2
* Resource action: child_rsc1:3 monitor on node1
* Resource action: child_rsc1:4 monitor on node2
* Resource action: child_rsc1:4 monitor on node1
* Pseudo action: rsc1_pre_notify_start_0
* Pseudo action: rsc1_confirmed-pre_notify_start_0
* Pseudo action: rsc1_start_0
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Resource action: child_rsc1:2 start on node1
* Resource action: child_rsc1:3 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_post_notify_running_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-post_notify_running_0
* Pseudo action: rsc1_pre_notify_promote_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-pre_notify_promote_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:0 promote on node1
* Resource action: child_rsc1:3 promote on node2
* Pseudo action: rsc1_promoted_0
* Pseudo action: rsc1_post_notify_promoted_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-post_notify_promoted_0
* Resource action: child_rsc1:0 monitor=11000 on node1
* Resource action: child_rsc1:1 monitor=1000 on node2
* Resource action: child_rsc1:2 monitor=1000 on node1
* Resource action: child_rsc1:3 monitor=11000 on node2
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Promoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:2 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:3 (ocf:heartbeat:apache): Promoted node2
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
diff --git a/cts/scheduler/summary/promoted-11.summary b/cts/scheduler/summary/promoted-11.summary
index 89fb85776d..47732fb9da 100644
--- a/cts/scheduler/summary/promoted-11.summary
+++ b/cts/scheduler/summary/promoted-11.summary
@@ -1,40 +1,40 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* simple-rsc (ocf:heartbeat:apache): Stopped
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Start simple-rsc ( node2 )
* Start child_rsc1:0 ( node1 )
* Promote child_rsc1:1 ( Stopped -> Promoted node2 )
Executing Cluster Transition:
* Resource action: simple-rsc monitor on node2
* Resource action: simple-rsc monitor on node1
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Pseudo action: rsc1_start_0
* Resource action: simple-rsc start on node2
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:1 promote on node2
* Pseudo action: rsc1_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* simple-rsc (ocf:heartbeat:apache): Started node2
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Promoted node2
diff --git a/cts/scheduler/summary/promoted-12.summary b/cts/scheduler/summary/promoted-12.summary
index 878d366883..9125a9aa28 100644
--- a/cts/scheduler/summary/promoted-12.summary
+++ b/cts/scheduler/summary/promoted-12.summary
@@ -1,33 +1,33 @@
Current cluster status:
* Node List:
* Online: [ sel3 sel4 ]
* Full List of Resources:
* Clone Set: ms-drbd0 [drbd0] (promotable):
* Promoted: [ sel3 ]
* Unpromoted: [ sel4 ]
- * Clone Set: ms-sf [sf] (promotable) (unique):
+ * Clone Set: ms-sf [sf] (promotable, unique):
* sf:0 (ocf:heartbeat:Stateful): Unpromoted sel3
* sf:1 (ocf:heartbeat:Stateful): Unpromoted sel4
* fs0 (ocf:heartbeat:Filesystem): Started sel3
Transition Summary:
* Promote sf:0 ( Unpromoted -> Promoted sel3 )
Executing Cluster Transition:
* Pseudo action: ms-sf_promote_0
* Resource action: sf:0 promote on sel3
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ sel3 sel4 ]
* Full List of Resources:
* Clone Set: ms-drbd0 [drbd0] (promotable):
* Promoted: [ sel3 ]
* Unpromoted: [ sel4 ]
- * Clone Set: ms-sf [sf] (promotable) (unique):
+ * Clone Set: ms-sf [sf] (promotable, unique):
* sf:0 (ocf:heartbeat:Stateful): Promoted sel3
* sf:1 (ocf:heartbeat:Stateful): Unpromoted sel4
* fs0 (ocf:heartbeat:Filesystem): Started sel3
diff --git a/cts/scheduler/summary/promoted-2.summary b/cts/scheduler/summary/promoted-2.summary
index 3258499fc8..9adf43ef1d 100644
--- a/cts/scheduler/summary/promoted-2.summary
+++ b/cts/scheduler/summary/promoted-2.summary
@@ -1,71 +1,71 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
* child_rsc1:2 (ocf:heartbeat:apache): Stopped
* child_rsc1:3 (ocf:heartbeat:apache): Stopped
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Promote child_rsc1:0 ( Stopped -> Promoted node1 )
* Start child_rsc1:1 ( node2 )
* Start child_rsc1:2 ( node1 )
* Promote child_rsc1:3 ( Stopped -> Promoted node2 )
Executing Cluster Transition:
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:2 monitor on node2
* Resource action: child_rsc1:2 monitor on node1
* Resource action: child_rsc1:3 monitor on node2
* Resource action: child_rsc1:3 monitor on node1
* Resource action: child_rsc1:4 monitor on node2
* Resource action: child_rsc1:4 monitor on node1
* Pseudo action: rsc1_pre_notify_start_0
* Pseudo action: rsc1_confirmed-pre_notify_start_0
* Pseudo action: rsc1_start_0
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Resource action: child_rsc1:2 start on node1
* Resource action: child_rsc1:3 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_post_notify_running_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-post_notify_running_0
* Pseudo action: rsc1_pre_notify_promote_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-pre_notify_promote_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:0 promote on node1
* Resource action: child_rsc1:3 promote on node2
* Pseudo action: rsc1_promoted_0
* Pseudo action: rsc1_post_notify_promoted_0
* Resource action: child_rsc1:0 notify on node1
* Resource action: child_rsc1:1 notify on node2
* Resource action: child_rsc1:2 notify on node1
* Resource action: child_rsc1:3 notify on node2
* Pseudo action: rsc1_confirmed-post_notify_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Promoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:2 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:3 (ocf:heartbeat:apache): Promoted node2
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
diff --git a/cts/scheduler/summary/promoted-3.summary b/cts/scheduler/summary/promoted-3.summary
index 161f51834a..08100f3e36 100644
--- a/cts/scheduler/summary/promoted-3.summary
+++ b/cts/scheduler/summary/promoted-3.summary
@@ -1,50 +1,50 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Stopped
* child_rsc1:1 (ocf:heartbeat:apache): Stopped
* child_rsc1:2 (ocf:heartbeat:apache): Stopped
* child_rsc1:3 (ocf:heartbeat:apache): Stopped
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
Transition Summary:
* Start child_rsc1:0 ( node1 )
* Promote child_rsc1:1 ( Stopped -> Promoted node2 )
* Start child_rsc1:2 ( node1 )
* Start child_rsc1:3 ( node2 )
Executing Cluster Transition:
* Resource action: child_rsc1:0 monitor on node2
* Resource action: child_rsc1:0 monitor on node1
* Resource action: child_rsc1:1 monitor on node2
* Resource action: child_rsc1:1 monitor on node1
* Resource action: child_rsc1:2 monitor on node2
* Resource action: child_rsc1:2 monitor on node1
* Resource action: child_rsc1:3 monitor on node2
* Resource action: child_rsc1:3 monitor on node1
* Resource action: child_rsc1:4 monitor on node2
* Resource action: child_rsc1:4 monitor on node1
* Pseudo action: rsc1_start_0
* Resource action: child_rsc1:0 start on node1
* Resource action: child_rsc1:1 start on node2
* Resource action: child_rsc1:2 start on node1
* Resource action: child_rsc1:3 start on node2
* Pseudo action: rsc1_running_0
* Pseudo action: rsc1_promote_0
* Resource action: child_rsc1:1 promote on node2
* Pseudo action: rsc1_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: rsc1 [child_rsc1] (promotable) (unique):
+ * Clone Set: rsc1 [child_rsc1] (promotable, unique):
* child_rsc1:0 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:1 (ocf:heartbeat:apache): Promoted node2
* child_rsc1:2 (ocf:heartbeat:apache): Unpromoted node1
* child_rsc1:3 (ocf:heartbeat:apache): Unpromoted node2
* child_rsc1:4 (ocf:heartbeat:apache): Stopped
diff --git a/cts/scheduler/summary/promoted-4.summary b/cts/scheduler/summary/promoted-4.summary
index 0dfe7c7263..2bcb25eaf1 100644
--- a/cts/scheduler/summary/promoted-4.summary
+++ b/cts/scheduler/summary/promoted-4.summary
@@ -1,94 +1,94 @@
Current cluster status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_child (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_child (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n01
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n01
* child_DoFencing:3 (stonith:ssh): Started c001n02
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
Transition Summary:
* Promote ocf_msdummy:0 ( Unpromoted -> Promoted c001n08 )
Executing Cluster Transition:
* Resource action: child_DoFencing:1 monitor on c001n08
* Resource action: child_DoFencing:1 monitor on c001n02
* Resource action: child_DoFencing:1 monitor on c001n01
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:2 monitor on c001n02
* Resource action: child_DoFencing:3 monitor on c001n08
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n01
* Resource action: ocf_msdummy:0 cancel=5000 on c001n08
* Resource action: ocf_msdummy:2 monitor on c001n08
* Resource action: ocf_msdummy:2 monitor on c001n03
* Resource action: ocf_msdummy:2 monitor on c001n02
* Resource action: ocf_msdummy:3 monitor on c001n03
* Resource action: ocf_msdummy:3 monitor on c001n02
* Resource action: ocf_msdummy:3 monitor on c001n01
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n02
* Resource action: ocf_msdummy:4 monitor on c001n01
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n03
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:6 monitor on c001n01
* Resource action: ocf_msdummy:7 monitor on c001n08
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n01
* Pseudo action: master_rsc_1_promote_0
* Resource action: ocf_msdummy:0 promote on c001n08
* Pseudo action: master_rsc_1_promoted_0
* Resource action: ocf_msdummy:0 monitor=6000 on c001n08
Revised Cluster Status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_child (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_child (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n01
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n01
* child_DoFencing:3 (stonith:ssh): Started c001n02
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
diff --git a/cts/scheduler/summary/promoted-5.summary b/cts/scheduler/summary/promoted-5.summary
index 00fa1c2154..8a2f1a232f 100644
--- a/cts/scheduler/summary/promoted-5.summary
+++ b/cts/scheduler/summary/promoted-5.summary
@@ -1,88 +1,88 @@
Current cluster status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_child (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_child (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n01
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n01
* child_DoFencing:3 (stonith:ssh): Started c001n02
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
Transition Summary:
Executing Cluster Transition:
* Resource action: child_DoFencing:1 monitor on c001n08
* Resource action: child_DoFencing:1 monitor on c001n02
* Resource action: child_DoFencing:1 monitor on c001n01
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:2 monitor on c001n02
* Resource action: child_DoFencing:3 monitor on c001n08
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n01
* Resource action: ocf_msdummy:2 monitor on c001n08
* Resource action: ocf_msdummy:2 monitor on c001n03
* Resource action: ocf_msdummy:2 monitor on c001n02
* Resource action: ocf_msdummy:3 monitor on c001n03
* Resource action: ocf_msdummy:3 monitor on c001n02
* Resource action: ocf_msdummy:3 monitor on c001n01
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n02
* Resource action: ocf_msdummy:4 monitor on c001n01
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n03
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:6 monitor on c001n01
* Resource action: ocf_msdummy:7 monitor on c001n08
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n01
Revised Cluster Status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_child (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_child (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n01
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n01
* child_DoFencing:3 (stonith:ssh): Started c001n02
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
diff --git a/cts/scheduler/summary/promoted-6.summary b/cts/scheduler/summary/promoted-6.summary
index 13c12dfc1c..2d9c953bfa 100644
--- a/cts/scheduler/summary/promoted-6.summary
+++ b/cts/scheduler/summary/promoted-6.summary
@@ -1,87 +1,87 @@
Current cluster status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n02
* child_DoFencing:2 (stonith:ssh): Started c001n03
* child_DoFencing:3 (stonith:ssh): Started c001n01
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
Transition Summary:
Executing Cluster Transition:
* Resource action: child_DoFencing:1 monitor on c001n08
* Resource action: child_DoFencing:1 monitor on c001n03
* Resource action: child_DoFencing:1 monitor on c001n01
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n01
* Resource action: child_DoFencing:3 monitor on c001n08
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n02
* Resource action: ocf_msdummy:2 monitor on c001n08
* Resource action: ocf_msdummy:2 monitor on c001n01
* Resource action: ocf_msdummy:3 monitor on c001n03
* Resource action: ocf_msdummy:3 monitor on c001n01
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n03
* Resource action: ocf_msdummy:4 monitor on c001n01
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:5 monitor on c001n01
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:6 monitor on c001n02
* Resource action: ocf_msdummy:7 monitor on c001n08
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n01 c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n08
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n08
* child_DoFencing:1 (stonith:ssh): Started c001n02
* child_DoFencing:2 (stonith:ssh): Started c001n03
* child_DoFencing:3 (stonith:ssh): Started c001n01
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n08
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01
diff --git a/cts/scheduler/summary/promoted-7.summary b/cts/scheduler/summary/promoted-7.summary
index 0602f95895..e43682c9d4 100644
--- a/cts/scheduler/summary/promoted-7.summary
+++ b/cts/scheduler/summary/promoted-7.summary
@@ -1,121 +1,121 @@
Current cluster status:
* Node List:
* Node c001n01: UNCLEAN (offline)
* Online: [ c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n03
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n02
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n01 (UNCLEAN)
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n01 (UNCLEAN)
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01 (UNCLEAN)
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
Transition Summary:
* Fence (reboot) c001n01 'peer is no longer part of the cluster'
* Move DcIPaddr ( c001n01 -> c001n03 )
* Move ocf_192.168.100.181 ( c001n03 -> c001n02 )
* Move heartbeat_192.168.100.182 ( c001n03 -> c001n02 )
* Move ocf_192.168.100.183 ( c001n03 -> c001n02 )
* Move lsb_dummy ( c001n02 -> c001n08 )
* Move rsc_c001n01 ( c001n01 -> c001n03 )
* Stop child_DoFencing:0 ( c001n01 ) due to node availability
* Stop ocf_msdummy:0 ( Promoted c001n01 ) due to node availability
* Stop ocf_msdummy:4 ( Unpromoted c001n01 ) due to node availability
Executing Cluster Transition:
* Pseudo action: group-1_stop_0
* Resource action: ocf_192.168.100.183 stop on c001n03
* Resource action: lsb_dummy stop on c001n02
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n02
* Pseudo action: DoFencing_stop_0
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n03
* Resource action: ocf_msdummy:4 monitor on c001n02
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n02
* Pseudo action: master_rsc_1_demote_0
* Fencing c001n01 (reboot)
* Pseudo action: DcIPaddr_stop_0
* Resource action: heartbeat_192.168.100.182 stop on c001n03
* Resource action: lsb_dummy start on c001n08
* Pseudo action: rsc_c001n01_stop_0
* Pseudo action: child_DoFencing:0_stop_0
* Pseudo action: DoFencing_stopped_0
* Pseudo action: ocf_msdummy:0_demote_0
* Pseudo action: master_rsc_1_demoted_0
* Pseudo action: master_rsc_1_stop_0
* Resource action: DcIPaddr start on c001n03
* Resource action: ocf_192.168.100.181 stop on c001n03
* Resource action: lsb_dummy monitor=5000 on c001n08
* Resource action: rsc_c001n01 start on c001n03
* Pseudo action: ocf_msdummy:0_stop_0
* Pseudo action: ocf_msdummy:4_stop_0
* Pseudo action: master_rsc_1_stopped_0
* Resource action: DcIPaddr monitor=5000 on c001n03
* Pseudo action: group-1_stopped_0
* Pseudo action: group-1_start_0
* Resource action: ocf_192.168.100.181 start on c001n02
* Resource action: heartbeat_192.168.100.182 start on c001n02
* Resource action: ocf_192.168.100.183 start on c001n02
* Resource action: rsc_c001n01 monitor=5000 on c001n03
* Pseudo action: group-1_running_0
* Resource action: ocf_192.168.100.181 monitor=5000 on c001n02
* Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02
* Resource action: ocf_192.168.100.183 monitor=5000 on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n08 ]
* OFFLINE: [ c001n01 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n03
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n08
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Stopped
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
diff --git a/cts/scheduler/summary/promoted-8.summary b/cts/scheduler/summary/promoted-8.summary
index 32417ff1ea..571eba6945 100644
--- a/cts/scheduler/summary/promoted-8.summary
+++ b/cts/scheduler/summary/promoted-8.summary
@@ -1,124 +1,124 @@
Current cluster status:
* Node List:
* Node c001n01: UNCLEAN (offline)
* Online: [ c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n03
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n02
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n01 (UNCLEAN)
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n01 (UNCLEAN)
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
Transition Summary:
* Fence (reboot) c001n01 'peer is no longer part of the cluster'
* Move DcIPaddr ( c001n01 -> c001n03 )
* Move ocf_192.168.100.181 ( c001n03 -> c001n02 )
* Move heartbeat_192.168.100.182 ( c001n03 -> c001n02 )
* Move ocf_192.168.100.183 ( c001n03 -> c001n02 )
* Move lsb_dummy ( c001n02 -> c001n08 )
* Move rsc_c001n01 ( c001n01 -> c001n03 )
* Stop child_DoFencing:0 ( c001n01 ) due to node availability
* Move ocf_msdummy:0 ( Promoted c001n01 -> Unpromoted c001n03 )
Executing Cluster Transition:
* Pseudo action: group-1_stop_0
* Resource action: ocf_192.168.100.183 stop on c001n03
* Resource action: lsb_dummy stop on c001n02
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n02
* Pseudo action: DoFencing_stop_0
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n03
* Resource action: ocf_msdummy:4 monitor on c001n02
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n03
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n02
* Pseudo action: master_rsc_1_demote_0
* Fencing c001n01 (reboot)
* Pseudo action: DcIPaddr_stop_0
* Resource action: heartbeat_192.168.100.182 stop on c001n03
* Resource action: lsb_dummy start on c001n08
* Pseudo action: rsc_c001n01_stop_0
* Pseudo action: child_DoFencing:0_stop_0
* Pseudo action: DoFencing_stopped_0
* Pseudo action: ocf_msdummy:0_demote_0
* Pseudo action: master_rsc_1_demoted_0
* Pseudo action: master_rsc_1_stop_0
* Resource action: DcIPaddr start on c001n03
* Resource action: ocf_192.168.100.181 stop on c001n03
* Resource action: lsb_dummy monitor=5000 on c001n08
* Resource action: rsc_c001n01 start on c001n03
* Pseudo action: ocf_msdummy:0_stop_0
* Pseudo action: master_rsc_1_stopped_0
* Pseudo action: master_rsc_1_start_0
* Resource action: DcIPaddr monitor=5000 on c001n03
* Pseudo action: group-1_stopped_0
* Pseudo action: group-1_start_0
* Resource action: ocf_192.168.100.181 start on c001n02
* Resource action: heartbeat_192.168.100.182 start on c001n02
* Resource action: ocf_192.168.100.183 start on c001n02
* Resource action: rsc_c001n01 monitor=5000 on c001n03
* Resource action: ocf_msdummy:0 start on c001n03
* Pseudo action: master_rsc_1_running_0
* Pseudo action: group-1_running_0
* Resource action: ocf_192.168.100.181 monitor=5000 on c001n02
* Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02
* Resource action: ocf_192.168.100.183 monitor=5000 on c001n02
* Resource action: ocf_msdummy:0 monitor=5000 on c001n03
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n08 ]
* OFFLINE: [ c001n01 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n03
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n08
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Stopped
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
diff --git a/cts/scheduler/summary/promoted-9.summary b/cts/scheduler/summary/promoted-9.summary
index 2da56a62db..7dfdbbda99 100644
--- a/cts/scheduler/summary/promoted-9.summary
+++ b/cts/scheduler/summary/promoted-9.summary
@@ -1,100 +1,100 @@
Current cluster status:
* Node List:
* Node sgi2: UNCLEAN (offline)
* Node test02: UNCLEAN (offline)
* Online: [ ibm1 va1 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_127.0.0.11 (ocf:heartbeat:IPaddr): Stopped
* heartbeat_127.0.0.12 (ocf:heartbeat:IPaddr): Stopped
* ocf_127.0.0.13 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib64/heartbeat/cts/LSBDummy): Stopped
* rsc_sgi2 (ocf:heartbeat:IPaddr): Stopped
* rsc_ibm1 (ocf:heartbeat:IPaddr): Stopped
* rsc_va1 (ocf:heartbeat:IPaddr): Stopped
* rsc_test02 (ocf:heartbeat:IPaddr): Stopped
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started va1
* child_DoFencing:1 (stonith:ssh): Started ibm1
* child_DoFencing:2 (stonith:ssh): Stopped
* child_DoFencing:3 (stonith:ssh): Stopped
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
Transition Summary:
* Start DcIPaddr ( va1 ) due to no quorum (blocked)
* Start ocf_127.0.0.11 ( va1 ) due to no quorum (blocked)
* Start heartbeat_127.0.0.12 ( va1 ) due to no quorum (blocked)
* Start ocf_127.0.0.13 ( va1 ) due to no quorum (blocked)
* Start lsb_dummy ( va1 ) due to no quorum (blocked)
* Start rsc_sgi2 ( va1 ) due to no quorum (blocked)
* Start rsc_ibm1 ( va1 ) due to no quorum (blocked)
* Start rsc_va1 ( va1 ) due to no quorum (blocked)
* Start rsc_test02 ( va1 ) due to no quorum (blocked)
* Stop child_DoFencing:1 ( ibm1 ) due to node availability
* Promote ocf_msdummy:0 ( Stopped -> Promoted va1 ) blocked
* Start ocf_msdummy:1 ( va1 ) due to no quorum (blocked)
Executing Cluster Transition:
* Resource action: child_DoFencing:1 monitor on va1
* Resource action: child_DoFencing:2 monitor on va1
* Resource action: child_DoFencing:2 monitor on ibm1
* Resource action: child_DoFencing:3 monitor on va1
* Resource action: child_DoFencing:3 monitor on ibm1
* Pseudo action: DoFencing_stop_0
* Resource action: ocf_msdummy:2 monitor on va1
* Resource action: ocf_msdummy:2 monitor on ibm1
* Resource action: ocf_msdummy:3 monitor on va1
* Resource action: ocf_msdummy:3 monitor on ibm1
* Resource action: ocf_msdummy:4 monitor on va1
* Resource action: ocf_msdummy:4 monitor on ibm1
* Resource action: ocf_msdummy:5 monitor on va1
* Resource action: ocf_msdummy:5 monitor on ibm1
* Resource action: ocf_msdummy:6 monitor on va1
* Resource action: ocf_msdummy:6 monitor on ibm1
* Resource action: ocf_msdummy:7 monitor on va1
* Resource action: ocf_msdummy:7 monitor on ibm1
* Resource action: child_DoFencing:1 stop on ibm1
* Pseudo action: DoFencing_stopped_0
* Cluster action: do_shutdown on ibm1
Revised Cluster Status:
* Node List:
* Node sgi2: UNCLEAN (offline)
* Node test02: UNCLEAN (offline)
* Online: [ ibm1 va1 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_127.0.0.11 (ocf:heartbeat:IPaddr): Stopped
* heartbeat_127.0.0.12 (ocf:heartbeat:IPaddr): Stopped
* ocf_127.0.0.13 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib64/heartbeat/cts/LSBDummy): Stopped
* rsc_sgi2 (ocf:heartbeat:IPaddr): Stopped
* rsc_ibm1 (ocf:heartbeat:IPaddr): Stopped
* rsc_va1 (ocf:heartbeat:IPaddr): Stopped
* rsc_test02 (ocf:heartbeat:IPaddr): Stopped
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started va1
* child_DoFencing:1 (stonith:ssh): Stopped
* child_DoFencing:2 (stonith:ssh): Stopped
* child_DoFencing:3 (stonith:ssh): Stopped
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped
diff --git a/cts/scheduler/summary/promoted-asymmetrical-order.summary b/cts/scheduler/summary/promoted-asymmetrical-order.summary
index e10568e898..1e49b3084b 100644
--- a/cts/scheduler/summary/promoted-asymmetrical-order.summary
+++ b/cts/scheduler/summary/promoted-asymmetrical-order.summary
@@ -1,37 +1,37 @@
2 of 4 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: ms1 [rsc1] (promotable) (disabled):
+ * Clone Set: ms1 [rsc1] (promotable, disabled):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
* Clone Set: ms2 [rsc2] (promotable):
* Promoted: [ node2 ]
* Unpromoted: [ node1 ]
Transition Summary:
* Stop rsc1:0 ( Promoted node1 ) due to node availability
* Stop rsc1:1 ( Unpromoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms1_demote_0
* Resource action: rsc1:0 demote on node1
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Resource action: rsc1:0 stop on node1
* Resource action: rsc1:1 stop on node2
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Clone Set: ms1 [rsc1] (promotable) (disabled):
+ * Clone Set: ms1 [rsc1] (promotable, disabled):
* Stopped (disabled): [ node1 node2 ]
* Clone Set: ms2 [rsc2] (promotable):
* Promoted: [ node2 ]
* Unpromoted: [ node1 ]
diff --git a/cts/scheduler/summary/promoted-failed-demote-2.summary b/cts/scheduler/summary/promoted-failed-demote-2.summary
index c8504e9e1d..453b5b7c9b 100644
--- a/cts/scheduler/summary/promoted-failed-demote-2.summary
+++ b/cts/scheduler/summary/promoted-failed-demote-2.summary
@@ -1,47 +1,47 @@
Current cluster status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
Transition Summary:
* Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
* Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a )
* Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a )
Executing Cluster Transition:
* Resource action: stateful-1:1 cancel=20000 on dl380g5a
* Resource action: stateful-2:1 cancel=20000 on dl380g5a
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-1:0 stop on dl380g5b
* Pseudo action: group:0_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-1:1 promote on dl380g5a
* Resource action: stateful-2:1 promote on dl380g5a
* Pseudo action: group:1_promoted_0
* Resource action: stateful-1:1 monitor=10000 on dl380g5a
* Resource action: stateful-2:1 monitor=10000 on dl380g5a
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Stopped
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
diff --git a/cts/scheduler/summary/promoted-failed-demote.summary b/cts/scheduler/summary/promoted-failed-demote.summary
index f071025528..732fba89c7 100644
--- a/cts/scheduler/summary/promoted-failed-demote.summary
+++ b/cts/scheduler/summary/promoted-failed-demote.summary
@@ -1,64 +1,64 @@
Current cluster status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
Transition Summary:
* Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
* Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a )
* Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a )
Executing Cluster Transition:
* Resource action: stateful-1:1 cancel=20000 on dl380g5a
* Resource action: stateful-2:1 cancel=20000 on dl380g5a
* Pseudo action: ms-sf_pre_notify_stop_0
* Resource action: stateful-1:0 notify on dl380g5b
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-pre_notify_stop_0
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-1:0 stop on dl380g5b
* Pseudo action: group:0_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_post_notify_stopped_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-post_notify_stopped_0
* Pseudo action: ms-sf_pre_notify_promote_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-pre_notify_promote_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-1:1 promote on dl380g5a
* Resource action: stateful-2:1 promote on dl380g5a
* Pseudo action: group:1_promoted_0
* Pseudo action: ms-sf_promoted_0
* Pseudo action: ms-sf_post_notify_promoted_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-post_notify_promoted_0
* Resource action: stateful-1:1 monitor=10000 on dl380g5a
* Resource action: stateful-2:1 monitor=10000 on dl380g5a
Revised Cluster Status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
- * Clone Set: ms-sf [group] (promotable) (unique):
+ * Clone Set: ms-sf [group] (promotable, unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Stopped
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
diff --git a/cts/scheduler/summary/promoted-group.summary b/cts/scheduler/summary/promoted-group.summary
index f06047c34f..594d1b50d4 100644
--- a/cts/scheduler/summary/promoted-group.summary
+++ b/cts/scheduler/summary/promoted-group.summary
@@ -1,37 +1,37 @@
Current cluster status:
* Node List:
* Online: [ rh44-1 rh44-2 ]
* Full List of Resources:
* Resource Group: test:
* resource_1 (ocf:heartbeat:IPaddr): Started rh44-1
- * Clone Set: ms-sf [grp_ms_sf] (promotable) (unique):
+ * Clone Set: ms-sf [grp_ms_sf] (promotable, unique):
* Resource Group: grp_ms_sf:0:
* master_slave_Stateful:0 (ocf:heartbeat:Stateful): Unpromoted rh44-2
* Resource Group: grp_ms_sf:1:
* master_slave_Stateful:1 (ocf:heartbeat:Stateful): Unpromoted rh44-1
Transition Summary:
* Promote master_slave_Stateful:1 ( Unpromoted -> Promoted rh44-1 )
Executing Cluster Transition:
* Resource action: master_slave_Stateful:1 cancel=5000 on rh44-1
* Pseudo action: ms-sf_promote_0
* Pseudo action: grp_ms_sf:1_promote_0
* Resource action: master_slave_Stateful:1 promote on rh44-1
* Pseudo action: grp_ms_sf:1_promoted_0
* Resource action: master_slave_Stateful:1 monitor=6000 on rh44-1
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ rh44-1 rh44-2 ]
* Full List of Resources:
* Resource Group: test:
* resource_1 (ocf:heartbeat:IPaddr): Started rh44-1
- * Clone Set: ms-sf [grp_ms_sf] (promotable) (unique):
+ * Clone Set: ms-sf [grp_ms_sf] (promotable, unique):
* Resource Group: grp_ms_sf:0:
* master_slave_Stateful:0 (ocf:heartbeat:Stateful): Unpromoted rh44-2
* Resource Group: grp_ms_sf:1:
* master_slave_Stateful:1 (ocf:heartbeat:Stateful): Promoted rh44-1
diff --git a/cts/scheduler/summary/promoted-reattach.summary b/cts/scheduler/summary/promoted-reattach.summary
index cf089d9324..8f07251f2e 100644
--- a/cts/scheduler/summary/promoted-reattach.summary
+++ b/cts/scheduler/summary/promoted-reattach.summary
@@ -1,34 +1,34 @@
Current cluster status:
* Node List:
* Online: [ dktest1 dktest2 ]
* Full List of Resources:
- * Clone Set: ms-drbd1 [drbd1] (promotable) (unmanaged):
+ * Clone Set: ms-drbd1 [drbd1] (promotable, unmanaged):
* drbd1 (ocf:heartbeat:drbd): Promoted dktest1 (unmanaged)
* drbd1 (ocf:heartbeat:drbd): Unpromoted dktest2 (unmanaged)
* Resource Group: apache (unmanaged):
* apache-vip (ocf:heartbeat:IPaddr2): Started dktest1 (unmanaged)
* mount (ocf:heartbeat:Filesystem): Started dktest1 (unmanaged)
* webserver (ocf:heartbeat:apache): Started dktest1 (unmanaged)
Transition Summary:
Executing Cluster Transition:
* Resource action: drbd1:0 monitor=10000 on dktest1
* Resource action: drbd1:0 monitor=11000 on dktest2
* Resource action: apache-vip monitor=60000 on dktest1
* Resource action: mount monitor=10000 on dktest1
* Resource action: webserver monitor=30000 on dktest1
Revised Cluster Status:
* Node List:
* Online: [ dktest1 dktest2 ]
* Full List of Resources:
- * Clone Set: ms-drbd1 [drbd1] (promotable) (unmanaged):
+ * Clone Set: ms-drbd1 [drbd1] (promotable, unmanaged):
* drbd1 (ocf:heartbeat:drbd): Promoted dktest1 (unmanaged)
* drbd1 (ocf:heartbeat:drbd): Unpromoted dktest2 (unmanaged)
* Resource Group: apache (unmanaged):
* apache-vip (ocf:heartbeat:IPaddr2): Started dktest1 (unmanaged)
* mount (ocf:heartbeat:Filesystem): Started dktest1 (unmanaged)
* webserver (ocf:heartbeat:apache): Started dktest1 (unmanaged)
diff --git a/cts/scheduler/summary/promoted-unmanaged-monitor.summary b/cts/scheduler/summary/promoted-unmanaged-monitor.summary
index 2b96429fad..3c5b39aa17 100644
--- a/cts/scheduler/summary/promoted-unmanaged-monitor.summary
+++ b/cts/scheduler/summary/promoted-unmanaged-monitor.summary
@@ -1,69 +1,69 @@
Current cluster status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild] (unmanaged):
* Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Resource Group: group-1 (unmanaged):
* r192.168.122.112 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* r192.168.122.113 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* r192.168.122.114 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 (unmanaged)
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 (unmanaged)
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-3 (unmanaged)
* migrator (ocf:pacemaker:Dummy): Started pcmk-4 (unmanaged)
* Clone Set: Connectivity [ping-1] (unmanaged):
* ping-1 (ocf:pacemaker:ping): Started pcmk-2 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-3 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-4 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-1 (unmanaged)
- * Clone Set: master-1 [stateful-1] (promotable) (unmanaged):
+ * Clone Set: master-1 [stateful-1] (promotable, unmanaged):
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-2 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Promoted pcmk-3 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-4 (unmanaged)
* Stopped: [ pcmk-1 ]
Transition Summary:
Executing Cluster Transition:
* Resource action: lsb-dummy monitor=5000 on pcmk-3
* Resource action: migrator monitor=10000 on pcmk-4
* Resource action: ping-1:0 monitor=60000 on pcmk-2
* Resource action: ping-1:0 monitor=60000 on pcmk-3
* Resource action: ping-1:0 monitor=60000 on pcmk-4
* Resource action: ping-1:0 monitor=60000 on pcmk-1
* Resource action: stateful-1:0 monitor=15000 on pcmk-2
* Resource action: stateful-1:0 monitor on pcmk-1
* Resource action: stateful-1:0 monitor=16000 on pcmk-3
* Resource action: stateful-1:0 monitor=15000 on pcmk-4
Revised Cluster Status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild] (unmanaged):
* Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Resource Group: group-1 (unmanaged):
* r192.168.122.112 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* r192.168.122.113 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* r192.168.122.114 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 (unmanaged)
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 (unmanaged)
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-3 (unmanaged)
* migrator (ocf:pacemaker:Dummy): Started pcmk-4 (unmanaged)
* Clone Set: Connectivity [ping-1] (unmanaged):
* ping-1 (ocf:pacemaker:ping): Started pcmk-2 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-3 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-4 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-1 (unmanaged)
- * Clone Set: master-1 [stateful-1] (promotable) (unmanaged):
+ * Clone Set: master-1 [stateful-1] (promotable, unmanaged):
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-2 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Promoted pcmk-3 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-4 (unmanaged)
* Stopped: [ pcmk-1 ]
diff --git a/cts/scheduler/summary/rec-node-13.summary b/cts/scheduler/summary/rec-node-13.summary
index 68210542c3..72c8e42736 100644
--- a/cts/scheduler/summary/rec-node-13.summary
+++ b/cts/scheduler/summary/rec-node-13.summary
@@ -1,80 +1,80 @@
Current cluster status:
* Node List:
* Node c001n04: UNCLEAN (online)
* Online: [ c001n02 c001n06 c001n07 ]
* OFFLINE: [ c001n03 c001n05 ]
* Full List of Resources:
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n02 c001n06 c001n07 ]
* Stopped: [ c001n03 c001n04 c001n05 ]
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n06
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n02
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): FAILED c001n04
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:8 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:9 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:10 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:11 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
Transition Summary:
* Fence (reboot) c001n04 'ocf_msdummy:6 failed there'
* Stop ocf_msdummy:6 ( Unpromoted c001n04 ) due to node availability
Executing Cluster Transition:
* Fencing c001n04 (reboot)
* Pseudo action: master_rsc_1_stop_0
* Pseudo action: ocf_msdummy:6_stop_0
* Pseudo action: master_rsc_1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n06 c001n07 ]
* OFFLINE: [ c001n03 c001n04 c001n05 ]
* Full List of Resources:
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n02 c001n06 c001n07 ]
* Stopped: [ c001n03 c001n04 c001n05 ]
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n06
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n02
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:8 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:9 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:10 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:11 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
diff --git a/cts/scheduler/summary/rsc-maintenance.summary b/cts/scheduler/summary/rsc-maintenance.summary
index fed1d61e19..0b9d57ed2a 100644
--- a/cts/scheduler/summary/rsc-maintenance.summary
+++ b/cts/scheduler/summary/rsc-maintenance.summary
@@ -1,31 +1,31 @@
2 of 4 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Resource Group: group1 (unmanaged) (disabled):
+ * Resource Group: group1 (unmanaged, disabled):
* rsc1 (ocf:pacemaker:Dummy): Started node1 (disabled, unmanaged)
* rsc2 (ocf:pacemaker:Dummy): Started node1 (disabled, unmanaged)
* Resource Group: group2:
* rsc3 (ocf:pacemaker:Dummy): Started node2
* rsc4 (ocf:pacemaker:Dummy): Started node2
Transition Summary:
Executing Cluster Transition:
* Resource action: rsc1 cancel=10000 on node1
* Resource action: rsc2 cancel=10000 on node1
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
- * Resource Group: group1 (unmanaged) (disabled):
+ * Resource Group: group1 (unmanaged, disabled):
* rsc1 (ocf:pacemaker:Dummy): Started node1 (disabled, unmanaged)
* rsc2 (ocf:pacemaker:Dummy): Started node1 (disabled, unmanaged)
* Resource Group: group2:
* rsc3 (ocf:pacemaker:Dummy): Started node2
* rsc4 (ocf:pacemaker:Dummy): Started node2
diff --git a/cts/scheduler/summary/stonith-0.summary b/cts/scheduler/summary/stonith-0.summary
index 5b829bf06d..f9745bd642 100644
--- a/cts/scheduler/summary/stonith-0.summary
+++ b/cts/scheduler/summary/stonith-0.summary
@@ -1,111 +1,111 @@
Current cluster status:
* Node List:
* Node c001n03: UNCLEAN (online)
* Node c001n05: UNCLEAN (online)
* Online: [ c001n02 c001n04 c001n06 c001n07 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started [ c001n03 c001n05 ]
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n03
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): FAILED [ c001n03 c001n05 ]
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n04
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n05
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n02 c001n04 c001n06 c001n07 c001n08 ]
* Stopped: [ c001n03 c001n05 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n02
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:8 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:9 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:10 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n04
* ocf_msdummy:11 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n04
* ocf_msdummy:12 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:13 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
Transition Summary:
* Fence (reboot) c001n05 'ocf_192.168.100.183 failed there'
* Fence (reboot) c001n03 'ocf_192.168.100.183 failed there'
* Move ocf_192.168.100.181 ( c001n03 -> c001n02 )
* Move heartbeat_192.168.100.182 ( c001n03 -> c001n02 )
* Recover ocf_192.168.100.183 ( c001n03 -> c001n02 )
* Move rsc_c001n05 ( c001n05 -> c001n07 )
* Move rsc_c001n07 ( c001n03 -> c001n07 )
Executing Cluster Transition:
* Resource action: child_DoFencing:4 monitor=20000 on c001n08
* Fencing c001n05 (reboot)
* Fencing c001n03 (reboot)
* Pseudo action: group-1_stop_0
* Pseudo action: ocf_192.168.100.183_stop_0
* Pseudo action: ocf_192.168.100.183_stop_0
* Pseudo action: rsc_c001n05_stop_0
* Pseudo action: rsc_c001n07_stop_0
* Pseudo action: heartbeat_192.168.100.182_stop_0
* Resource action: rsc_c001n05 start on c001n07
* Resource action: rsc_c001n07 start on c001n07
* Pseudo action: ocf_192.168.100.181_stop_0
* Pseudo action: ocf_192.168.100.181_stop_0
* Resource action: rsc_c001n05 monitor=5000 on c001n07
* Resource action: rsc_c001n07 monitor=5000 on c001n07
* Pseudo action: group-1_stopped_0
* Pseudo action: group-1_start_0
* Resource action: ocf_192.168.100.181 start on c001n02
* Resource action: heartbeat_192.168.100.182 start on c001n02
* Resource action: ocf_192.168.100.183 start on c001n02
* Pseudo action: group-1_running_0
* Resource action: ocf_192.168.100.181 monitor=5000 on c001n02
* Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02
* Resource action: ocf_192.168.100.183 monitor=5000 on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n04 c001n06 c001n07 c001n08 ]
* OFFLINE: [ c001n03 c001n05 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Stopped
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n04
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n04 (ocf:heartbeat:IPaddr): Started c001n04
* rsc_c001n05 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n06 (ocf:heartbeat:IPaddr): Started c001n06
* rsc_c001n07 (ocf:heartbeat:IPaddr): Started c001n07
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ c001n02 c001n04 c001n06 c001n07 c001n08 ]
* Stopped: [ c001n03 c001n05 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n02
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n07
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:8 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:9 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:10 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n04
* ocf_msdummy:11 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n04
* ocf_msdummy:12 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
* ocf_msdummy:13 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n06
diff --git a/cts/scheduler/summary/stonith-1.summary b/cts/scheduler/summary/stonith-1.summary
index 31d2f57ce8..29b979cacc 100644
--- a/cts/scheduler/summary/stonith-1.summary
+++ b/cts/scheduler/summary/stonith-1.summary
@@ -1,113 +1,113 @@
Current cluster status:
* Node List:
* Node sles-3: UNCLEAN (offline)
* Online: [ sles-1 sles-2 sles-4 ]
* Full List of Resources:
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.182 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.183 (ocf:heartbeat:IPaddr): Stopped
* lsb_dummy (lsb:/usr/lib64/heartbeat/cts/LSBDummy): Started sles-2
* migrator (ocf:heartbeat:Dummy): Started sles-3 (UNCLEAN)
* rsc_sles-1 (ocf:heartbeat:IPaddr): Started sles-1
* rsc_sles-2 (ocf:heartbeat:IPaddr): Started sles-2
* rsc_sles-3 (ocf:heartbeat:IPaddr): Started sles-3 (UNCLEAN)
* rsc_sles-4 (ocf:heartbeat:IPaddr): Started sles-4
* Clone Set: DoFencing [child_DoFencing]:
* child_DoFencing (stonith:external/vmware): Started sles-3 (UNCLEAN)
* Started: [ sles-1 sles-2 ]
* Stopped: [ sles-4 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Unpromoted sles-3 (UNCLEAN)
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Unpromoted sles-3 (UNCLEAN)
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
Transition Summary:
* Fence (reboot) sles-3 'peer is no longer part of the cluster'
* Start r192.168.100.183 ( sles-1 )
* Move migrator ( sles-3 -> sles-4 )
* Move rsc_sles-3 ( sles-3 -> sles-4 )
* Move child_DoFencing:2 ( sles-3 -> sles-4 )
* Start ocf_msdummy:0 ( sles-4 )
* Start ocf_msdummy:1 ( sles-1 )
* Move ocf_msdummy:2 ( sles-3 -> sles-2 Unpromoted )
* Start ocf_msdummy:3 ( sles-4 )
* Start ocf_msdummy:4 ( sles-1 )
* Move ocf_msdummy:5 ( sles-3 -> sles-2 Unpromoted )
Executing Cluster Transition:
* Pseudo action: group-1_start_0
* Resource action: r192.168.100.182 monitor=5000 on sles-1
* Resource action: lsb_dummy monitor=5000 on sles-2
* Resource action: rsc_sles-2 monitor=5000 on sles-2
* Resource action: rsc_sles-4 monitor=5000 on sles-4
* Pseudo action: DoFencing_stop_0
* Fencing sles-3 (reboot)
* Resource action: r192.168.100.183 start on sles-1
* Pseudo action: migrator_stop_0
* Pseudo action: rsc_sles-3_stop_0
* Pseudo action: child_DoFencing:2_stop_0
* Pseudo action: DoFencing_stopped_0
* Pseudo action: DoFencing_start_0
* Pseudo action: master_rsc_1_stop_0
* Pseudo action: group-1_running_0
* Resource action: r192.168.100.183 monitor=5000 on sles-1
* Resource action: migrator start on sles-4
* Resource action: rsc_sles-3 start on sles-4
* Resource action: child_DoFencing:2 start on sles-4
* Pseudo action: DoFencing_running_0
* Pseudo action: ocf_msdummy:2_stop_0
* Pseudo action: ocf_msdummy:5_stop_0
* Pseudo action: master_rsc_1_stopped_0
* Pseudo action: master_rsc_1_start_0
* Resource action: migrator monitor=10000 on sles-4
* Resource action: rsc_sles-3 monitor=5000 on sles-4
* Resource action: child_DoFencing:2 monitor=60000 on sles-4
* Resource action: ocf_msdummy:0 start on sles-4
* Resource action: ocf_msdummy:1 start on sles-1
* Resource action: ocf_msdummy:2 start on sles-2
* Resource action: ocf_msdummy:3 start on sles-4
* Resource action: ocf_msdummy:4 start on sles-1
* Resource action: ocf_msdummy:5 start on sles-2
* Pseudo action: master_rsc_1_running_0
* Resource action: ocf_msdummy:0 monitor=5000 on sles-4
* Resource action: ocf_msdummy:1 monitor=5000 on sles-1
* Resource action: ocf_msdummy:2 monitor=5000 on sles-2
* Resource action: ocf_msdummy:3 monitor=5000 on sles-4
* Resource action: ocf_msdummy:4 monitor=5000 on sles-1
* Resource action: ocf_msdummy:5 monitor=5000 on sles-2
Revised Cluster Status:
* Node List:
* Online: [ sles-1 sles-2 sles-4 ]
* OFFLINE: [ sles-3 ]
* Full List of Resources:
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.182 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.183 (ocf:heartbeat:IPaddr): Started sles-1
* lsb_dummy (lsb:/usr/lib64/heartbeat/cts/LSBDummy): Started sles-2
* migrator (ocf:heartbeat:Dummy): Started sles-4
* rsc_sles-1 (ocf:heartbeat:IPaddr): Started sles-1
* rsc_sles-2 (ocf:heartbeat:IPaddr): Started sles-2
* rsc_sles-3 (ocf:heartbeat:IPaddr): Started sles-4
* rsc_sles-4 (ocf:heartbeat:IPaddr): Started sles-4
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ sles-1 sles-2 sles-4 ]
* Stopped: [ sles-3 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
diff --git a/cts/scheduler/summary/stonith-2.summary b/cts/scheduler/summary/stonith-2.summary
index 9fd7c65e59..c6f657193b 100644
--- a/cts/scheduler/summary/stonith-2.summary
+++ b/cts/scheduler/summary/stonith-2.summary
@@ -1,78 +1,78 @@
Current cluster status:
* Node List:
* Node sles-5: UNCLEAN (offline)
* Online: [ sles-1 sles-2 sles-3 sles-4 sles-6 ]
* Full List of Resources:
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.182 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.183 (ocf:heartbeat:IPaddr): Started sles-1
* lsb_dummy (lsb:/usr/share/heartbeat/cts/LSBDummy): Started sles-2
* migrator (ocf:heartbeat:Dummy): Started sles-3
* rsc_sles-1 (ocf:heartbeat:IPaddr): Started sles-1
* rsc_sles-2 (ocf:heartbeat:IPaddr): Started sles-2
* rsc_sles-3 (ocf:heartbeat:IPaddr): Started sles-3
* rsc_sles-4 (ocf:heartbeat:IPaddr): Started sles-4
* rsc_sles-5 (ocf:heartbeat:IPaddr): Stopped
* rsc_sles-6 (ocf:heartbeat:IPaddr): Started sles-6
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ sles-1 sles-2 sles-3 sles-4 sles-6 ]
* Stopped: [ sles-5 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Unpromoted sles-3
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Unpromoted sles-6
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Unpromoted sles-6
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Unpromoted sles-3
Transition Summary:
* Fence (reboot) sles-5 'peer is no longer part of the cluster'
* Start rsc_sles-5 ( sles-6 )
Executing Cluster Transition:
* Fencing sles-5 (reboot)
* Resource action: rsc_sles-5 start on sles-6
* Resource action: rsc_sles-5 monitor=5000 on sles-6
Revised Cluster Status:
* Node List:
* Online: [ sles-1 sles-2 sles-3 sles-4 sles-6 ]
* OFFLINE: [ sles-5 ]
* Full List of Resources:
* Resource Group: group-1:
* r192.168.100.181 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.182 (ocf:heartbeat:IPaddr): Started sles-1
* r192.168.100.183 (ocf:heartbeat:IPaddr): Started sles-1
* lsb_dummy (lsb:/usr/share/heartbeat/cts/LSBDummy): Started sles-2
* migrator (ocf:heartbeat:Dummy): Started sles-3
* rsc_sles-1 (ocf:heartbeat:IPaddr): Started sles-1
* rsc_sles-2 (ocf:heartbeat:IPaddr): Started sles-2
* rsc_sles-3 (ocf:heartbeat:IPaddr): Started sles-3
* rsc_sles-4 (ocf:heartbeat:IPaddr): Started sles-4
* rsc_sles-5 (ocf:heartbeat:IPaddr): Started sles-6
* rsc_sles-6 (ocf:heartbeat:IPaddr): Started sles-6
* Clone Set: DoFencing [child_DoFencing]:
* Started: [ sles-1 sles-2 sles-3 sles-4 sles-6 ]
* Stopped: [ sles-5 ]
- * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
+ * Clone Set: master_rsc_1 [ocf_msdummy] (promotable, unique):
* ocf_msdummy:0 (ocf:heartbeat:Stateful): Unpromoted sles-3
* ocf_msdummy:1 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:2 (ocf:heartbeat:Stateful): Unpromoted sles-4
* ocf_msdummy:3 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:4 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:5 (ocf:heartbeat:Stateful): Unpromoted sles-1
* ocf_msdummy:6 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:7 (ocf:heartbeat:Stateful): Stopped
* ocf_msdummy:8 (ocf:heartbeat:Stateful): Unpromoted sles-6
* ocf_msdummy:9 (ocf:heartbeat:Stateful): Unpromoted sles-6
* ocf_msdummy:10 (ocf:heartbeat:Stateful): Unpromoted sles-2
* ocf_msdummy:11 (ocf:heartbeat:Stateful): Unpromoted sles-3
diff --git a/cts/scheduler/summary/unmanaged-promoted.summary b/cts/scheduler/summary/unmanaged-promoted.summary
index bdaac99618..33c0a4127f 100644
--- a/cts/scheduler/summary/unmanaged-promoted.summary
+++ b/cts/scheduler/summary/unmanaged-promoted.summary
@@ -1,63 +1,63 @@
Current cluster status:
* Node List:
* Online: [ pcmk-1 pcmk-2 ]
* OFFLINE: [ pcmk-3 pcmk-4 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild] (unmanaged):
* FencingChild (stonith:fence_xvm): Started pcmk-2 (unmanaged)
* FencingChild (stonith:fence_xvm): Started pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
* Resource Group: group-1 (unmanaged):
* r192.168.122.126 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* r192.168.122.127 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* r192.168.122.128 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 (unmanaged)
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 (unmanaged)
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2 (unmanaged)
* migrator (ocf:pacemaker:Dummy): Started pcmk-4 (unmanaged)
* Clone Set: Connectivity [ping-1] (unmanaged):
* ping-1 (ocf:pacemaker:ping): Started pcmk-2 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
- * Clone Set: master-1 [stateful-1] (promotable) (unmanaged):
+ * Clone Set: master-1 [stateful-1] (promotable, unmanaged):
* stateful-1 (ocf:pacemaker:Stateful): Promoted pcmk-2 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
Transition Summary:
Executing Cluster Transition:
* Cluster action: do_shutdown on pcmk-2
* Cluster action: do_shutdown on pcmk-1
Revised Cluster Status:
* Node List:
* Online: [ pcmk-1 pcmk-2 ]
* OFFLINE: [ pcmk-3 pcmk-4 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild] (unmanaged):
* FencingChild (stonith:fence_xvm): Started pcmk-2 (unmanaged)
* FencingChild (stonith:fence_xvm): Started pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
* Resource Group: group-1 (unmanaged):
* r192.168.122.126 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* r192.168.122.127 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* r192.168.122.128 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 (unmanaged)
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 (unmanaged)
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 (unmanaged)
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 (unmanaged)
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2 (unmanaged)
* migrator (ocf:pacemaker:Dummy): Started pcmk-4 (unmanaged)
* Clone Set: Connectivity [ping-1] (unmanaged):
* ping-1 (ocf:pacemaker:ping): Started pcmk-2 (unmanaged)
* ping-1 (ocf:pacemaker:ping): Started pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
- * Clone Set: master-1 [stateful-1] (promotable) (unmanaged):
+ * Clone Set: master-1 [stateful-1] (promotable, unmanaged):
* stateful-1 (ocf:pacemaker:Stateful): Promoted pcmk-2 (unmanaged)
* stateful-1 (ocf:pacemaker:Stateful): Unpromoted pcmk-1 (unmanaged)
* Stopped: [ pcmk-3 pcmk-4 ]
diff --git a/lib/pengine/clone.c b/lib/pengine/clone.c
index b8cfc497b3..7f4d0744ba 100644
--- a/lib/pengine/clone.c
+++ b/lib/pengine/clone.c
@@ -1,1017 +1,1039 @@
/*
* Copyright 2004-2021 the Pacemaker project contributors
*
* The version control history for this file may have further details.
*
* This source code is licensed under the GNU Lesser General Public License
* version 2.1 or later (LGPLv2.1+) WITHOUT ANY WARRANTY.
*/
#include
#include
#include
#include
#include
#include
#include
#include
#define VARIANT_CLONE 1
#include "./variant.h"
#ifdef PCMK__COMPAT_2_0
#define PROMOTED_INSTANCES RSC_ROLE_PROMOTED_LEGACY_S "s"
#define UNPROMOTED_INSTANCES RSC_ROLE_UNPROMOTED_LEGACY_S "s"
#else
#define PROMOTED_INSTANCES RSC_ROLE_PROMOTED_S
#define UNPROMOTED_INSTANCES RSC_ROLE_UNPROMOTED_S
#endif
static void
clone_header(pcmk__output_t *out, int *rc, pe_resource_t *rsc, clone_variant_data_t *clone_data)
{
- PCMK__OUTPUT_LIST_HEADER(out, FALSE, *rc, "Clone Set: %s [%s]%s%s%s%s",
+ char *attrs = NULL;
+ size_t len = 0;
+
+ if (pcmk_is_set(rsc->flags, pe_rsc_promotable)) {
+ pcmk__add_separated_word(&attrs, &len, "promotable", ", ");
+ }
+
+ if (pcmk_is_set(rsc->flags, pe_rsc_unique)) {
+ pcmk__add_separated_word(&attrs, &len, "unique", ", ");
+ }
+
+ if (!pcmk_is_set(rsc->flags, pe_rsc_managed)) {
+ pcmk__add_separated_word(&attrs, &len, "unmanaged", ", ");
+ }
+
+ if (pe__resource_is_disabled(rsc)) {
+ pcmk__add_separated_word(&attrs, &len, "disabled", ", ");
+ }
+
+ if (attrs) {
+ PCMK__OUTPUT_LIST_HEADER(out, FALSE, *rc, "Clone Set: %s [%s] (%s)",
rsc->id, ID(clone_data->xml_obj_child),
- pcmk_is_set(rsc->flags, pe_rsc_promotable) ? " (promotable)" : "",
- pcmk_is_set(rsc->flags, pe_rsc_unique) ? " (unique)" : "",
- pcmk_is_set(rsc->flags, pe_rsc_managed) ? "" : " (unmanaged)",
- pe__resource_is_disabled(rsc) ? " (disabled)" : "");
+ attrs);
+ free(attrs);
+ } else {
+ PCMK__OUTPUT_LIST_HEADER(out, FALSE, *rc, "Clone Set: %s [%s]",
+ rsc->id, ID(clone_data->xml_obj_child))
+ }
}
void
pe__force_anon(const char *standard, pe_resource_t *rsc, const char *rid,
pe_working_set_t *data_set)
{
if (pe_rsc_is_clone(rsc)) {
clone_variant_data_t *clone_data = NULL;
get_clone_variant_data(clone_data, rsc);
pe_warn("Ignoring " XML_RSC_ATTR_UNIQUE " for %s because %s resources "
"such as %s can be used only as anonymous clones",
rsc->id, standard, rid);
clone_data->clone_node_max = 1;
clone_data->clone_max = QB_MIN(clone_data->clone_max,
g_list_length(data_set->nodes));
}
}
pe_resource_t *
find_clone_instance(pe_resource_t * rsc, const char *sub_id, pe_working_set_t * data_set)
{
char *child_id = NULL;
pe_resource_t *child = NULL;
const char *child_base = NULL;
clone_variant_data_t *clone_data = NULL;
get_clone_variant_data(clone_data, rsc);
child_base = ID(clone_data->xml_obj_child);
child_id = crm_strdup_printf("%s:%s", child_base, sub_id);
child = pe_find_resource(rsc->children, child_id);
free(child_id);
return child;
}
pe_resource_t *
pe__create_clone_child(pe_resource_t *rsc, pe_working_set_t *data_set)
{
gboolean as_orphan = FALSE;
char *inc_num = NULL;
char *inc_max = NULL;
pe_resource_t *child_rsc = NULL;
xmlNode *child_copy = NULL;
clone_variant_data_t *clone_data = NULL;
get_clone_variant_data(clone_data, rsc);
CRM_CHECK(clone_data->xml_obj_child != NULL, return FALSE);
if (clone_data->total_clones >= clone_data->clone_max) {
// If we've already used all available instances, this is an orphan
as_orphan = TRUE;
}
// Allocate instance numbers in numerical order (starting at 0)
inc_num = pcmk__itoa(clone_data->total_clones);
inc_max = pcmk__itoa(clone_data->clone_max);
child_copy = copy_xml(clone_data->xml_obj_child);
crm_xml_add(child_copy, XML_RSC_ATTR_INCARNATION, inc_num);
if (common_unpack(child_copy, &child_rsc, rsc, data_set) == FALSE) {
pe_err("Failed unpacking resource %s", crm_element_value(child_copy, XML_ATTR_ID));
child_rsc = NULL;
goto bail;
}
/* child_rsc->globally_unique = rsc->globally_unique; */
CRM_ASSERT(child_rsc);
clone_data->total_clones += 1;
pe_rsc_trace(child_rsc, "Setting clone attributes for: %s", child_rsc->id);
rsc->children = g_list_append(rsc->children, child_rsc);
if (as_orphan) {
pe__set_resource_flags_recursive(child_rsc, pe_rsc_orphan);
}
add_hash_param(child_rsc->meta, XML_RSC_ATTR_INCARNATION_MAX, inc_max);
pe_rsc_trace(rsc, "Added %s instance %s", rsc->id, child_rsc->id);
bail:
free(inc_num);
free(inc_max);
return child_rsc;
}
gboolean
clone_unpack(pe_resource_t * rsc, pe_working_set_t * data_set)
{
int lpc = 0;
xmlNode *a_child = NULL;
xmlNode *xml_obj = rsc->xml;
clone_variant_data_t *clone_data = NULL;
const char *ordered = g_hash_table_lookup(rsc->meta, XML_RSC_ATTR_ORDERED);
const char *max_clones = g_hash_table_lookup(rsc->meta, XML_RSC_ATTR_INCARNATION_MAX);
const char *max_clones_node = g_hash_table_lookup(rsc->meta, XML_RSC_ATTR_INCARNATION_NODEMAX);
pe_rsc_trace(rsc, "Processing resource %s...", rsc->id);
clone_data = calloc(1, sizeof(clone_variant_data_t));
rsc->variant_opaque = clone_data;
if (pcmk_is_set(rsc->flags, pe_rsc_promotable)) {
const char *promoted_max = NULL;
const char *promoted_node_max = NULL;
promoted_max = g_hash_table_lookup(rsc->meta,
XML_RSC_ATTR_PROMOTED_MAX);
if (promoted_max == NULL) {
// @COMPAT deprecated since 2.0.0
promoted_max = g_hash_table_lookup(rsc->meta,
PCMK_XE_PROMOTED_MAX_LEGACY);
}
promoted_node_max = g_hash_table_lookup(rsc->meta,
XML_RSC_ATTR_PROMOTED_NODEMAX);
if (promoted_node_max == NULL) {
// @COMPAT deprecated since 2.0.0
promoted_node_max = g_hash_table_lookup(rsc->meta,
PCMK_XE_PROMOTED_NODE_MAX_LEGACY);
}
// Use 1 as default but 0 for minimum and invalid
if (promoted_max == NULL) {
clone_data->promoted_max = 1;
} else {
pcmk__scan_min_int(promoted_max, &(clone_data->promoted_max), 0);
}
// Use 1 as default but 0 for minimum and invalid
if (promoted_node_max == NULL) {
clone_data->promoted_node_max = 1;
} else {
pcmk__scan_min_int(promoted_node_max,
&(clone_data->promoted_node_max), 0);
}
}
// Implied by calloc()
/* clone_data->xml_obj_child = NULL; */
// Use 1 as default but 0 for minimum and invalid
if (max_clones_node == NULL) {
clone_data->clone_node_max = 1;
} else {
pcmk__scan_min_int(max_clones_node, &(clone_data->clone_node_max), 0);
}
/* Use number of nodes (but always at least 1, which is handy for crm_verify
* for a CIB without nodes) as default, but 0 for minimum and invalid
*/
if (max_clones == NULL) {
clone_data->clone_max = QB_MAX(1, g_list_length(data_set->nodes));
} else {
pcmk__scan_min_int(max_clones, &(clone_data->clone_max), 0);
}
clone_data->ordered = crm_is_true(ordered);
if ((rsc->flags & pe_rsc_unique) == 0 && clone_data->clone_node_max > 1) {
pcmk__config_err("Ignoring " XML_RSC_ATTR_PROMOTED_MAX " for %s "
"because anonymous clones support only one instance "
"per node", rsc->id);
clone_data->clone_node_max = 1;
}
pe_rsc_trace(rsc, "Options for %s", rsc->id);
pe_rsc_trace(rsc, "\tClone max: %d", clone_data->clone_max);
pe_rsc_trace(rsc, "\tClone node max: %d", clone_data->clone_node_max);
pe_rsc_trace(rsc, "\tClone is unique: %s",
pe__rsc_bool_str(rsc, pe_rsc_unique));
pe_rsc_trace(rsc, "\tClone is promotable: %s",
pe__rsc_bool_str(rsc, pe_rsc_promotable));
// Clones may contain a single group or primitive
for (a_child = pcmk__xe_first_child(xml_obj); a_child != NULL;
a_child = pcmk__xe_next(a_child)) {
if (pcmk__str_any_of((const char *)a_child->name, XML_CIB_TAG_RESOURCE, XML_CIB_TAG_GROUP, NULL)) {
clone_data->xml_obj_child = a_child;
break;
}
}
if (clone_data->xml_obj_child == NULL) {
pcmk__config_err("%s has nothing to clone", rsc->id);
return FALSE;
}
/*
* Make clones ever so slightly sticky by default
*
* This helps ensure clone instances are not shuffled around the cluster
* for no benefit in situations when pre-allocation is not appropriate
*/
if (g_hash_table_lookup(rsc->meta, XML_RSC_ATTR_STICKINESS) == NULL) {
add_hash_param(rsc->meta, XML_RSC_ATTR_STICKINESS, "1");
}
/* This ensures that the globally-unique value always exists for children to
* inherit when being unpacked, as well as in resource agents' environment.
*/
add_hash_param(rsc->meta, XML_RSC_ATTR_UNIQUE,
pe__rsc_bool_str(rsc, pe_rsc_unique));
if (clone_data->clone_max <= 0) {
/* Create one child instance so that unpack_find_resource() will hook up
* any orphans up to the parent correctly.
*/
if (pe__create_clone_child(rsc, data_set) == NULL) {
return FALSE;
}
} else {
// Create a child instance for each available instance number
for (lpc = 0; lpc < clone_data->clone_max; lpc++) {
if (pe__create_clone_child(rsc, data_set) == NULL) {
return FALSE;
}
}
}
pe_rsc_trace(rsc, "Added %d children to resource %s...", clone_data->clone_max, rsc->id);
return TRUE;
}
gboolean
clone_active(pe_resource_t * rsc, gboolean all)
{
GList *gIter = rsc->children;
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
gboolean child_active = child_rsc->fns->active(child_rsc, all);
if (all == FALSE && child_active) {
return TRUE;
} else if (all && child_active == FALSE) {
return FALSE;
}
}
if (all) {
return TRUE;
} else {
return FALSE;
}
}
static void
short_print(char *list, const char *prefix, const char *type, const char *suffix, long options, void *print_data)
{
if(suffix == NULL) {
suffix = "";
}
if (list) {
if (options & pe_print_html) {
status_print("");
}
status_print("%s%s: [ %s ]%s", prefix, type, list, suffix);
if (options & pe_print_html) {
status_print("\n");
} else if (options & pe_print_suppres_nl) {
/* nothing */
} else if ((options & pe_print_printf) || (options & pe_print_ncurses)) {
status_print("\n");
}
}
}
static const char *
configured_role_str(pe_resource_t * rsc)
{
const char *target_role = g_hash_table_lookup(rsc->meta,
XML_RSC_ATTR_TARGET_ROLE);
if ((target_role == NULL) && rsc->children && rsc->children->data) {
target_role = g_hash_table_lookup(((pe_resource_t*)rsc->children->data)->meta,
XML_RSC_ATTR_TARGET_ROLE);
}
return target_role;
}
static enum rsc_role_e
configured_role(pe_resource_t * rsc)
{
const char *target_role = configured_role_str(rsc);
if (target_role) {
return text2role(target_role);
}
return RSC_ROLE_UNKNOWN;
}
static void
clone_print_xml(pe_resource_t * rsc, const char *pre_text, long options, void *print_data)
{
char *child_text = crm_strdup_printf("%s ", pre_text);
const char *target_role = configured_role_str(rsc);
GList *gIter = rsc->children;
status_print("%sid);
status_print("multi_state=\"%s\" ",
pe__rsc_bool_str(rsc, pe_rsc_promotable));
status_print("unique=\"%s\" ", pe__rsc_bool_str(rsc, pe_rsc_unique));
status_print("managed=\"%s\" ", pe__rsc_bool_str(rsc, pe_rsc_managed));
status_print("failed=\"%s\" ", pe__rsc_bool_str(rsc, pe_rsc_failed));
status_print("failure_ignored=\"%s\" ",
pe__rsc_bool_str(rsc, pe_rsc_failure_ignored));
if (target_role) {
status_print("target_role=\"%s\" ", target_role);
}
status_print(">\n");
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
child_rsc->fns->print(child_rsc, child_text, options, print_data);
}
status_print("%s\n", pre_text);
free(child_text);
}
bool is_set_recursive(pe_resource_t * rsc, long long flag, bool any)
{
GList *gIter;
bool all = !any;
if (pcmk_is_set(rsc->flags, flag)) {
if(any) {
return TRUE;
}
} else if(all) {
return FALSE;
}
for (gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
if(is_set_recursive(gIter->data, flag, any)) {
if(any) {
return TRUE;
}
} else if(all) {
return FALSE;
}
}
if(all) {
return TRUE;
}
return FALSE;
}
void
clone_print(pe_resource_t * rsc, const char *pre_text, long options, void *print_data)
{
char *list_text = NULL;
char *child_text = NULL;
char *stopped_list = NULL;
size_t list_text_len = 0;
size_t stopped_list_len = 0;
GList *promoted_list = NULL;
GList *started_list = NULL;
GList *gIter = rsc->children;
clone_variant_data_t *clone_data = NULL;
int active_instances = 0;
if (pre_text == NULL) {
pre_text = " ";
}
if (options & pe_print_xml) {
clone_print_xml(rsc, pre_text, options, print_data);
return;
}
get_clone_variant_data(clone_data, rsc);
child_text = crm_strdup_printf("%s ", pre_text);
status_print("%sClone Set: %s [%s]%s%s%s",
pre_text ? pre_text : "", rsc->id, ID(clone_data->xml_obj_child),
pcmk_is_set(rsc->flags, pe_rsc_promotable)? " (promotable)" : "",
pcmk_is_set(rsc->flags, pe_rsc_unique)? " (unique)" : "",
pcmk_is_set(rsc->flags, pe_rsc_managed)? "" : " (unmanaged)");
if (options & pe_print_html) {
status_print("\n\n");
} else if ((options & pe_print_log) == 0) {
status_print("\n");
}
for (; gIter != NULL; gIter = gIter->next) {
gboolean print_full = FALSE;
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
gboolean partially_active = child_rsc->fns->active(child_rsc, FALSE);
if (options & pe_print_clone_details) {
print_full = TRUE;
}
if (pcmk_is_set(rsc->flags, pe_rsc_unique)) {
// Print individual instance when unique (except stopped orphans)
if (partially_active || !pcmk_is_set(rsc->flags, pe_rsc_orphan)) {
print_full = TRUE;
}
// Everything else in this block is for anonymous clones
} else if (pcmk_is_set(options, pe_print_pending)
&& (child_rsc->pending_task != NULL)
&& strcmp(child_rsc->pending_task, "probe")) {
// Print individual instance when non-probe action is pending
print_full = TRUE;
} else if (partially_active == FALSE) {
// List stopped instances when requested (except orphans)
if (!pcmk_is_set(child_rsc->flags, pe_rsc_orphan)
&& !pcmk_is_set(options, pe_print_clone_active)) {
pcmk__add_word(&stopped_list, &stopped_list_len, child_rsc->id);
}
} else if (is_set_recursive(child_rsc, pe_rsc_orphan, TRUE)
|| is_set_recursive(child_rsc, pe_rsc_managed, FALSE) == FALSE
|| is_set_recursive(child_rsc, pe_rsc_failed, TRUE)) {
// Print individual instance when active orphaned/unmanaged/failed
print_full = TRUE;
} else if (child_rsc->fns->active(child_rsc, TRUE)) {
// Instance of fully active anonymous clone
pe_node_t *location = child_rsc->fns->location(child_rsc, NULL, TRUE);
if (location) {
// Instance is active on a single node
enum rsc_role_e a_role = child_rsc->fns->state(child_rsc, TRUE);
if (location->details->online == FALSE && location->details->unclean) {
print_full = TRUE;
} else if (a_role > RSC_ROLE_UNPROMOTED) {
promoted_list = g_list_append(promoted_list, location);
} else {
started_list = g_list_append(started_list, location);
}
} else {
/* uncolocated group - bleh */
print_full = TRUE;
}
} else {
// Instance of partially active anonymous clone
print_full = TRUE;
}
if (print_full) {
if (options & pe_print_html) {
status_print("- \n");
}
child_rsc->fns->print(child_rsc, child_text, options, print_data);
if (options & pe_print_html) {
status_print("
\n");
}
}
}
/* Promoted */
promoted_list = g_list_sort(promoted_list, sort_node_uname);
for (gIter = promoted_list; gIter; gIter = gIter->next) {
pe_node_t *host = gIter->data;
pcmk__add_word(&list_text, &list_text_len, host->details->uname);
active_instances++;
}
short_print(list_text, child_text, PROMOTED_INSTANCES, NULL, options,
print_data);
g_list_free(promoted_list);
free(list_text);
list_text = NULL;
list_text_len = 0;
/* Started/Unpromoted */
started_list = g_list_sort(started_list, sort_node_uname);
for (gIter = started_list; gIter; gIter = gIter->next) {
pe_node_t *host = gIter->data;
pcmk__add_word(&list_text, &list_text_len, host->details->uname);
active_instances++;
}
if (pcmk_is_set(rsc->flags, pe_rsc_promotable)) {
enum rsc_role_e role = configured_role(rsc);
if (role == RSC_ROLE_UNPROMOTED) {
short_print(list_text, child_text,
UNPROMOTED_INSTANCES " (target-role)", NULL, options,
print_data);
} else {
short_print(list_text, child_text, UNPROMOTED_INSTANCES, NULL,
options, print_data);
}
} else {
short_print(list_text, child_text, "Started", NULL, options, print_data);
}
g_list_free(started_list);
free(list_text);
list_text = NULL;
list_text_len = 0;
if (!pcmk_is_set(options, pe_print_clone_active)) {
const char *state = "Stopped";
enum rsc_role_e role = configured_role(rsc);
if (role == RSC_ROLE_STOPPED) {
state = "Stopped (disabled)";
}
if (!pcmk_is_set(rsc->flags, pe_rsc_unique)
&& (clone_data->clone_max > active_instances)) {
GList *nIter;
GList *list = g_hash_table_get_values(rsc->allowed_nodes);
/* Custom stopped list for non-unique clones */
free(stopped_list);
stopped_list = NULL;
stopped_list_len = 0;
if (list == NULL) {
/* Clusters with symmetrical=false haven't calculated allowed_nodes yet
* If we've not probed for them yet, the Stopped list will be empty
*/
list = g_hash_table_get_values(rsc->known_on);
}
list = g_list_sort(list, sort_node_uname);
for (nIter = list; nIter != NULL; nIter = nIter->next) {
pe_node_t *node = (pe_node_t *)nIter->data;
if (pe_find_node(rsc->running_on, node->details->uname) == NULL) {
pcmk__add_word(&stopped_list, &stopped_list_len,
node->details->uname);
}
}
g_list_free(list);
}
short_print(stopped_list, child_text, state, NULL, options, print_data);
free(stopped_list);
}
if (options & pe_print_html) {
status_print("
\n");
}
free(child_text);
}
PCMK__OUTPUT_ARGS("clone", "unsigned int", "pe_resource_t *", "GList *", "GList *")
int
pe__clone_xml(pcmk__output_t *out, va_list args)
{
unsigned int show_opts = va_arg(args, unsigned int);
pe_resource_t *rsc = va_arg(args, pe_resource_t *);
GList *only_node = va_arg(args, GList *);
GList *only_rsc = va_arg(args, GList *);
GList *gIter = rsc->children;
int rc = pcmk_rc_no_output;
gboolean printed_header = FALSE;
gboolean print_everything = TRUE;
if (rsc->fns->is_filtered(rsc, only_rsc, TRUE)) {
return rc;
}
print_everything = pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none) ||
(strstr(rsc->id, ":") != NULL && pcmk__str_in_list(only_rsc, rsc->id, pcmk__str_none));
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (pcmk__rsc_filtered_by_node(child_rsc, only_node)) {
continue;
}
if (child_rsc->fns->is_filtered(child_rsc, only_rsc, print_everything)) {
continue;
}
if (!printed_header) {
printed_header = TRUE;
rc = pe__name_and_nvpairs_xml(out, true, "clone", 8,
"id", rsc->id,
"multi_state", pe__rsc_bool_str(rsc, pe_rsc_promotable),
"unique", pe__rsc_bool_str(rsc, pe_rsc_unique),
"managed", pe__rsc_bool_str(rsc, pe_rsc_managed),
"disabled", pcmk__btoa(pe__resource_is_disabled(rsc)),
"failed", pe__rsc_bool_str(rsc, pe_rsc_failed),
"failure_ignored", pe__rsc_bool_str(rsc, pe_rsc_failure_ignored),
"target_role", configured_role_str(rsc));
CRM_ASSERT(rc == pcmk_rc_ok);
}
out->message(out, crm_map_element_name(child_rsc->xml), show_opts,
child_rsc, only_node, only_rsc);
}
if (printed_header) {
pcmk__output_xml_pop_parent(out);
}
return rc;
}
PCMK__OUTPUT_ARGS("clone", "unsigned int", "pe_resource_t *", "GList *", "GList *")
int
pe__clone_default(pcmk__output_t *out, va_list args)
{
unsigned int show_opts = va_arg(args, unsigned int);
pe_resource_t *rsc = va_arg(args, pe_resource_t *);
GList *only_node = va_arg(args, GList *);
GList *only_rsc = va_arg(args, GList *);
char *list_text = NULL;
char *stopped_list = NULL;
size_t list_text_len = 0;
size_t stopped_list_len = 0;
GList *promoted_list = NULL;
GList *started_list = NULL;
GList *gIter = rsc->children;
clone_variant_data_t *clone_data = NULL;
int active_instances = 0;
int rc = pcmk_rc_no_output;
gboolean print_everything = TRUE;
get_clone_variant_data(clone_data, rsc);
if (rsc->fns->is_filtered(rsc, only_rsc, TRUE)) {
return rc;
}
print_everything = pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none) ||
(strstr(rsc->id, ":") != NULL && pcmk__str_in_list(only_rsc, rsc->id, pcmk__str_none));
for (; gIter != NULL; gIter = gIter->next) {
gboolean print_full = FALSE;
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
gboolean partially_active = child_rsc->fns->active(child_rsc, FALSE);
if (pcmk__rsc_filtered_by_node(child_rsc, only_node)) {
continue;
}
if (child_rsc->fns->is_filtered(child_rsc, only_rsc, print_everything)) {
continue;
}
if (pcmk_is_set(show_opts, pcmk_show_clone_detail)) {
print_full = TRUE;
}
if (pcmk_is_set(rsc->flags, pe_rsc_unique)) {
// Print individual instance when unique (except stopped orphans)
if (partially_active || !pcmk_is_set(rsc->flags, pe_rsc_orphan)) {
print_full = TRUE;
}
// Everything else in this block is for anonymous clones
} else if (pcmk_is_set(show_opts, pcmk_show_pending)
&& (child_rsc->pending_task != NULL)
&& strcmp(child_rsc->pending_task, "probe")) {
// Print individual instance when non-probe action is pending
print_full = TRUE;
} else if (partially_active == FALSE) {
// List stopped instances when requested (except orphans)
if (!pcmk_is_set(child_rsc->flags, pe_rsc_orphan)
&& pcmk_is_set(show_opts, pcmk_show_inactive_rscs)) {
pcmk__add_word(&stopped_list, &stopped_list_len, child_rsc->id);
}
} else if (is_set_recursive(child_rsc, pe_rsc_orphan, TRUE)
|| is_set_recursive(child_rsc, pe_rsc_managed, FALSE) == FALSE
|| is_set_recursive(child_rsc, pe_rsc_failed, TRUE)) {
// Print individual instance when active orphaned/unmanaged/failed
print_full = TRUE;
} else if (child_rsc->fns->active(child_rsc, TRUE)) {
// Instance of fully active anonymous clone
pe_node_t *location = child_rsc->fns->location(child_rsc, NULL, TRUE);
if (location) {
// Instance is active on a single node
enum rsc_role_e a_role = child_rsc->fns->state(child_rsc, TRUE);
if (location->details->online == FALSE && location->details->unclean) {
print_full = TRUE;
} else if (a_role > RSC_ROLE_UNPROMOTED) {
promoted_list = g_list_append(promoted_list, location);
} else {
started_list = g_list_append(started_list, location);
}
} else {
/* uncolocated group - bleh */
print_full = TRUE;
}
} else {
// Instance of partially active anonymous clone
print_full = TRUE;
}
if (print_full) {
GList *all = NULL;
clone_header(out, &rc, rsc, clone_data);
/* Print every resource that's a child of this clone. */
all = g_list_prepend(all, (gpointer) "*");
out->message(out, crm_map_element_name(child_rsc->xml), show_opts,
child_rsc, only_node, all);
g_list_free(all);
}
}
if (pcmk_is_set(show_opts, pcmk_show_clone_detail)) {
free(stopped_list);
PCMK__OUTPUT_LIST_FOOTER(out, rc);
return pcmk_rc_ok;
}
/* Promoted */
promoted_list = g_list_sort(promoted_list, sort_node_uname);
for (gIter = promoted_list; gIter; gIter = gIter->next) {
pe_node_t *host = gIter->data;
if (!pcmk__str_in_list(only_node, host->details->uname, pcmk__str_casei)) {
continue;
}
pcmk__add_word(&list_text, &list_text_len, host->details->uname);
active_instances++;
}
if (list_text != NULL) {
clone_header(out, &rc, rsc, clone_data);
out->list_item(out, NULL, PROMOTED_INSTANCES ": [ %s ]", list_text);
g_list_free(promoted_list);
free(list_text);
list_text = NULL;
list_text_len = 0;
}
/* Started/Unpromoted */
started_list = g_list_sort(started_list, sort_node_uname);
for (gIter = started_list; gIter; gIter = gIter->next) {
pe_node_t *host = gIter->data;
if (!pcmk__str_in_list(only_node, host->details->uname, pcmk__str_casei)) {
continue;
}
pcmk__add_word(&list_text, &list_text_len, host->details->uname);
active_instances++;
}
if (list_text != NULL) {
clone_header(out, &rc, rsc, clone_data);
if (pcmk_is_set(rsc->flags, pe_rsc_promotable)) {
enum rsc_role_e role = configured_role(rsc);
if (role == RSC_ROLE_UNPROMOTED) {
out->list_item(out, NULL,
UNPROMOTED_INSTANCES " (target-role): [ %s ]",
list_text);
} else {
out->list_item(out, NULL, UNPROMOTED_INSTANCES ": [ %s ]",
list_text);
}
} else {
out->list_item(out, NULL, "Started: [ %s ]", list_text);
}
g_list_free(started_list);
free(list_text);
list_text = NULL;
list_text_len = 0;
}
if (pcmk_is_set(show_opts, pcmk_show_inactive_rscs)) {
const char *state = "Stopped";
enum rsc_role_e role = configured_role(rsc);
if (role == RSC_ROLE_STOPPED) {
state = "Stopped (disabled)";
}
if (!pcmk_is_set(rsc->flags, pe_rsc_unique)
&& (clone_data->clone_max > active_instances)) {
GList *nIter;
GList *list = g_hash_table_get_values(rsc->allowed_nodes);
/* Custom stopped list for non-unique clones */
free(stopped_list);
stopped_list = NULL;
stopped_list_len = 0;
if (list == NULL) {
/* Clusters with symmetrical=false haven't calculated allowed_nodes yet
* If we've not probed for them yet, the Stopped list will be empty
*/
list = g_hash_table_get_values(rsc->known_on);
}
list = g_list_sort(list, sort_node_uname);
for (nIter = list; nIter != NULL; nIter = nIter->next) {
pe_node_t *node = (pe_node_t *)nIter->data;
if (pe_find_node(rsc->running_on, node->details->uname) == NULL &&
pcmk__str_in_list(only_node, node->details->uname, pcmk__str_casei)) {
pcmk__add_word(&stopped_list, &stopped_list_len,
node->details->uname);
}
}
g_list_free(list);
}
if (stopped_list != NULL) {
clone_header(out, &rc, rsc, clone_data);
out->list_item(out, NULL, "%s: [ %s ]", state, stopped_list);
free(stopped_list);
stopped_list_len = 0;
}
}
PCMK__OUTPUT_LIST_FOOTER(out, rc);
return rc;
}
void
clone_free(pe_resource_t * rsc)
{
clone_variant_data_t *clone_data = NULL;
get_clone_variant_data(clone_data, rsc);
pe_rsc_trace(rsc, "Freeing %s", rsc->id);
for (GList *gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
CRM_ASSERT(child_rsc);
pe_rsc_trace(child_rsc, "Freeing child %s", child_rsc->id);
free_xml(child_rsc->xml);
child_rsc->xml = NULL;
/* There could be a saved unexpanded xml */
free_xml(child_rsc->orig_xml);
child_rsc->orig_xml = NULL;
child_rsc->fns->free(child_rsc);
}
g_list_free(rsc->children);
if (clone_data) {
CRM_ASSERT(clone_data->demote_notify == NULL);
CRM_ASSERT(clone_data->stop_notify == NULL);
CRM_ASSERT(clone_data->start_notify == NULL);
CRM_ASSERT(clone_data->promote_notify == NULL);
}
common_free(rsc);
}
enum rsc_role_e
clone_resource_state(const pe_resource_t * rsc, gboolean current)
{
enum rsc_role_e clone_role = RSC_ROLE_UNKNOWN;
GList *gIter = rsc->children;
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
enum rsc_role_e a_role = child_rsc->fns->state(child_rsc, current);
if (a_role > clone_role) {
clone_role = a_role;
}
}
pe_rsc_trace(rsc, "%s role: %s", rsc->id, role2text(clone_role));
return clone_role;
}
/*!
* \internal
* \brief Check whether a clone has an instance for every node
*
* \param[in] rsc Clone to check
* \param[in] data_set Cluster state
*/
bool
pe__is_universal_clone(pe_resource_t *rsc,
pe_working_set_t *data_set)
{
if (pe_rsc_is_clone(rsc)) {
clone_variant_data_t *clone_data = NULL;
get_clone_variant_data(clone_data, rsc);
if (clone_data->clone_max == g_list_length(data_set->nodes)) {
return TRUE;
}
}
return FALSE;
}
gboolean
pe__clone_is_filtered(pe_resource_t *rsc, GList *only_rsc, gboolean check_parent)
{
gboolean passes = FALSE;
clone_variant_data_t *clone_data = NULL;
if (pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none)) {
passes = TRUE;
} else {
get_clone_variant_data(clone_data, rsc);
passes = pcmk__str_in_list(only_rsc, ID(clone_data->xml_obj_child), pcmk__str_none);
if (!passes) {
for (GList *gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (!child_rsc->fns->is_filtered(child_rsc, only_rsc, FALSE)) {
passes = TRUE;
break;
}
}
}
}
return !passes;
}
diff --git a/lib/pengine/group.c b/lib/pengine/group.c
index ae94fc7268..31e73d7eb2 100644
--- a/lib/pengine/group.c
+++ b/lib/pengine/group.c
@@ -1,428 +1,423 @@
/*
* Copyright 2004-2021 the Pacemaker project contributors
*
* The version control history for this file may have further details.
*
* This source code is licensed under the GNU Lesser General Public License
* version 2.1 or later (LGPLv2.1+) WITHOUT ANY WARRANTY.
*/
#include
#include
#include
#include
#include
#include
#include
#include
#include
#define VARIANT_GROUP 1
#include "./variant.h"
static int
inactive_resources(pe_resource_t *rsc)
{
int retval = 0;
for (GList *gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (!child_rsc->fns->active(child_rsc, TRUE)) {
retval++;
}
}
return retval;
}
+static void
+group_header(pcmk__output_t *out, int *rc, pe_resource_t *rsc, int n_inactive, bool show_inactive)
+{
+ char *attrs = NULL;
+ size_t len = 0;
+
+ if (n_inactive > 0 && !show_inactive) {
+ char *word = crm_strdup_printf("%d member%s inactive", n_inactive, pcmk__plural_s(n_inactive));
+ pcmk__add_separated_word(&attrs, &len, word, ", ");
+ free(word);
+ }
+
+ if (!pcmk_is_set(rsc->flags, pe_rsc_managed)) {
+ pcmk__add_separated_word(&attrs, &len, "unmanaged", ", ");
+ }
+
+ if (pe__resource_is_disabled(rsc)) {
+ pcmk__add_separated_word(&attrs, &len, "disabled", ", ");
+ }
+
+ if (attrs) {
+ PCMK__OUTPUT_LIST_HEADER(out, FALSE, *rc, "Resource Group: %s (%s)",
+ rsc->id, attrs);
+ free(attrs);
+ } else {
+ PCMK__OUTPUT_LIST_HEADER(out, FALSE, *rc, "Resource Group: %s", rsc->id);
+ }
+}
+
static bool
skip_child_rsc(pe_resource_t *rsc, pe_resource_t *child, gboolean parent_passes,
GList *only_rsc, unsigned int show_opts)
{
bool star_list = pcmk__list_of_1(only_rsc) &&
pcmk__str_eq("*", g_list_first(only_rsc)->data, pcmk__str_none);
bool child_filtered = child->fns->is_filtered(child, only_rsc, FALSE);
bool child_active = child->fns->active(child, FALSE);
bool show_inactive = pcmk_is_set(show_opts, pcmk_show_inactive_rscs);
/* If the resource is in only_rsc by name (so, ignoring "*") then allow
* it regardless of if it's active or not.
*/
if (!star_list && !child_filtered) {
return false;
} else if (!child_filtered && (child_active || show_inactive)) {
return false;
} else if (parent_passes && (child_active || show_inactive)) {
return false;
}
return true;
}
gboolean
group_unpack(pe_resource_t * rsc, pe_working_set_t * data_set)
{
xmlNode *xml_obj = rsc->xml;
xmlNode *xml_native_rsc = NULL;
group_variant_data_t *group_data = NULL;
const char *group_ordered = g_hash_table_lookup(rsc->meta, XML_RSC_ATTR_ORDERED);
const char *group_colocated = g_hash_table_lookup(rsc->meta, "collocated");
const char *clone_id = NULL;
pe_rsc_trace(rsc, "Processing resource %s...", rsc->id);
group_data = calloc(1, sizeof(group_variant_data_t));
group_data->num_children = 0;
group_data->first_child = NULL;
group_data->last_child = NULL;
rsc->variant_opaque = group_data;
// We don't actually need the null checks but it speeds up the common case
if ((group_ordered == NULL)
|| (crm_str_to_boolean(group_ordered, &(group_data->ordered)) < 0)) {
group_data->ordered = TRUE;
}
if ((group_colocated == NULL)
|| (crm_str_to_boolean(group_colocated, &(group_data->colocated)) < 0)) {
group_data->colocated = TRUE;
}
clone_id = crm_element_value(rsc->xml, XML_RSC_ATTR_INCARNATION);
for (xml_native_rsc = pcmk__xe_first_child(xml_obj); xml_native_rsc != NULL;
xml_native_rsc = pcmk__xe_next(xml_native_rsc)) {
if (pcmk__str_eq((const char *)xml_native_rsc->name,
XML_CIB_TAG_RESOURCE, pcmk__str_none)) {
pe_resource_t *new_rsc = NULL;
crm_xml_add(xml_native_rsc, XML_RSC_ATTR_INCARNATION, clone_id);
if (common_unpack(xml_native_rsc, &new_rsc, rsc, data_set) == FALSE) {
pe_err("Failed unpacking resource %s", crm_element_value(xml_obj, XML_ATTR_ID));
if (new_rsc != NULL && new_rsc->fns != NULL) {
new_rsc->fns->free(new_rsc);
}
continue;
}
group_data->num_children++;
rsc->children = g_list_append(rsc->children, new_rsc);
if (group_data->first_child == NULL) {
group_data->first_child = new_rsc;
}
group_data->last_child = new_rsc;
pe_rsc_trace(rsc, "Added %s member %s", rsc->id, new_rsc->id);
}
}
if (group_data->num_children == 0) {
pcmk__config_warn("Group %s does not have any children", rsc->id);
return TRUE; // Allow empty groups, children can be added later
}
pe_rsc_trace(rsc, "Added %d children to resource %s...", group_data->num_children, rsc->id);
return TRUE;
}
gboolean
group_active(pe_resource_t * rsc, gboolean all)
{
gboolean c_all = TRUE;
gboolean c_any = FALSE;
GList *gIter = rsc->children;
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (child_rsc->fns->active(child_rsc, all)) {
c_any = TRUE;
} else {
c_all = FALSE;
}
}
if (c_any == FALSE) {
return FALSE;
} else if (all && c_all == FALSE) {
return FALSE;
}
return TRUE;
}
static void
group_print_xml(pe_resource_t * rsc, const char *pre_text, long options, void *print_data)
{
GList *gIter = rsc->children;
char *child_text = crm_strdup_printf("%s ", pre_text);
status_print("%sid);
status_print("number_resources=\"%d\" ", g_list_length(rsc->children));
status_print(">\n");
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
child_rsc->fns->print(child_rsc, child_text, options, print_data);
}
status_print("%s\n", pre_text);
free(child_text);
}
void
group_print(pe_resource_t * rsc, const char *pre_text, long options, void *print_data)
{
char *child_text = NULL;
GList *gIter = rsc->children;
if (pre_text == NULL) {
pre_text = " ";
}
if (options & pe_print_xml) {
group_print_xml(rsc, pre_text, options, print_data);
return;
}
child_text = crm_strdup_printf("%s ", pre_text);
status_print("%sResource Group: %s", pre_text ? pre_text : "", rsc->id);
if (options & pe_print_html) {
status_print("\n\n");
} else if ((options & pe_print_log) == 0) {
status_print("\n");
}
if (options & pe_print_brief) {
print_rscs_brief(rsc->children, child_text, options, print_data, TRUE);
} else {
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (options & pe_print_html) {
status_print("- \n");
}
child_rsc->fns->print(child_rsc, child_text, options, print_data);
if (options & pe_print_html) {
status_print("
\n");
}
}
}
if (options & pe_print_html) {
status_print("
\n");
}
free(child_text);
}
PCMK__OUTPUT_ARGS("group", "unsigned int", "pe_resource_t *", "GList *", "GList *")
int
pe__group_xml(pcmk__output_t *out, va_list args)
{
unsigned int show_opts = va_arg(args, unsigned int);
pe_resource_t *rsc = va_arg(args, pe_resource_t *);
GList *only_node = va_arg(args, GList *);
GList *only_rsc = va_arg(args, GList *);
GList *gIter = rsc->children;
char *count = pcmk__itoa(g_list_length(gIter));
int rc = pcmk_rc_no_output;
gboolean parent_passes = pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none) ||
(strstr(rsc->id, ":") != NULL && pcmk__str_in_list(only_rsc, rsc->id, pcmk__str_none));
if (rsc->fns->is_filtered(rsc, only_rsc, TRUE)) {
free(count);
return rc;
}
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (skip_child_rsc(rsc, child_rsc, parent_passes, only_rsc, show_opts)) {
continue;
}
if (rc == pcmk_rc_no_output) {
rc = pe__name_and_nvpairs_xml(out, true, "group", 4
, "id", rsc->id
, "number_resources", count
, "managed", pe__rsc_bool_str(rsc, pe_rsc_managed)
, "disabled", pcmk__btoa(pe__resource_is_disabled(rsc)));
free(count);
CRM_ASSERT(rc == pcmk_rc_ok);
}
out->message(out, crm_map_element_name(child_rsc->xml), show_opts, child_rsc,
only_node, only_rsc);
}
if (rc == pcmk_rc_ok) {
pcmk__output_xml_pop_parent(out);
}
return rc;
}
PCMK__OUTPUT_ARGS("group", "unsigned int", "pe_resource_t *", "GList *", "GList *")
int
pe__group_default(pcmk__output_t *out, va_list args)
{
unsigned int show_opts = va_arg(args, unsigned int);
pe_resource_t *rsc = va_arg(args, pe_resource_t *);
GList *only_node = va_arg(args, GList *);
GList *only_rsc = va_arg(args, GList *);
int rc = pcmk_rc_no_output;
gboolean parent_passes = pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none) ||
(strstr(rsc->id, ":") != NULL && pcmk__str_in_list(only_rsc, rsc->id, pcmk__str_none));
gboolean active = rsc->fns->active(rsc, TRUE);
gboolean partially_active = rsc->fns->active(rsc, FALSE);
if (rsc->fns->is_filtered(rsc, only_rsc, TRUE)) {
return rc;
}
if (pcmk_is_set(show_opts, pcmk_show_brief)) {
GList *rscs = pe__filter_rsc_list(rsc->children, only_rsc);
if (rscs != NULL) {
- char *s = NULL;
-
- if (!active && partially_active && !pcmk_is_set(show_opts, pcmk_show_inactive_rscs)) {
- int n_active = inactive_resources(rsc);
-
- if (n_active > 0) {
- s = crm_strdup_printf(" (%d member%s inactive)", n_active, pcmk__plural_s(n_active));
- }
- }
-
- out->begin_list(out, NULL, NULL, "Resource Group: %s%s%s%s", rsc->id,
- s ? s : "",
- pcmk_is_set(rsc->flags, pe_rsc_managed) ? "" : " (unmanaged)",
- pe__resource_is_disabled(rsc) ? " (disabled)" : "");
-
+ group_header(out, &rc, rsc, !active && partially_active ? inactive_resources(rsc) : 0,
+ pcmk_is_set(show_opts, pcmk_show_inactive_rscs));
pe__rscs_brief_output(out, rscs, show_opts | pcmk_show_inactive_rscs);
rc = pcmk_rc_ok;
g_list_free(rscs);
-
- if (s) {
- free(s);
- }
}
} else {
for (GList *gIter = rsc->children; gIter; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
- char *s = NULL;
if (skip_child_rsc(rsc, child_rsc, parent_passes, only_rsc, show_opts)) {
continue;
}
- if (!active && partially_active && !pcmk_is_set(show_opts, pcmk_show_inactive_rscs)) {
- int n_active = inactive_resources(rsc);
-
- if (n_active > 0) {
- s = crm_strdup_printf(" (%d member%s inactive)", n_active, pcmk__plural_s(n_active));
- }
- }
-
-
- PCMK__OUTPUT_LIST_HEADER(out, FALSE, rc, "Resource Group: %s%s%s%s", rsc->id,
- s ? s : "",
- pcmk_is_set(rsc->flags, pe_rsc_managed) ? "" : " (unmanaged)",
- pe__resource_is_disabled(rsc) ? " (disabled)" : "");
-
+ group_header(out, &rc, rsc, !active && partially_active ? inactive_resources(rsc) : 0,
+ pcmk_is_set(show_opts, pcmk_show_inactive_rscs));
out->message(out, crm_map_element_name(child_rsc->xml), show_opts,
child_rsc, only_node, only_rsc);
-
- if (s) {
- free(s);
- }
}
}
PCMK__OUTPUT_LIST_FOOTER(out, rc);
return rc;
}
void
group_free(pe_resource_t * rsc)
{
CRM_CHECK(rsc != NULL, return);
pe_rsc_trace(rsc, "Freeing %s", rsc->id);
for (GList *gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
CRM_ASSERT(child_rsc);
pe_rsc_trace(child_rsc, "Freeing child %s", child_rsc->id);
child_rsc->fns->free(child_rsc);
}
pe_rsc_trace(rsc, "Freeing child list");
g_list_free(rsc->children);
common_free(rsc);
}
enum rsc_role_e
group_resource_state(const pe_resource_t * rsc, gboolean current)
{
enum rsc_role_e group_role = RSC_ROLE_UNKNOWN;
GList *gIter = rsc->children;
for (; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
enum rsc_role_e role = child_rsc->fns->state(child_rsc, current);
if (role > group_role) {
group_role = role;
}
}
pe_rsc_trace(rsc, "%s role: %s", rsc->id, role2text(group_role));
return group_role;
}
gboolean
pe__group_is_filtered(pe_resource_t *rsc, GList *only_rsc, gboolean check_parent)
{
gboolean passes = FALSE;
if (check_parent && pcmk__str_in_list(only_rsc, rsc_printable_id(uber_parent(rsc)), pcmk__str_none)) {
passes = TRUE;
} else if (pcmk__str_in_list(only_rsc, rsc_printable_id(rsc), pcmk__str_none)) {
passes = TRUE;
} else if (strstr(rsc->id, ":") != NULL && pcmk__str_in_list(only_rsc, rsc->id, pcmk__str_none)) {
passes = TRUE;
} else {
for (GList *gIter = rsc->children; gIter != NULL; gIter = gIter->next) {
pe_resource_t *child_rsc = (pe_resource_t *) gIter->data;
if (!child_rsc->fns->is_filtered(child_rsc, only_rsc, FALSE)) {
passes = TRUE;
break;
}
}
}
return !passes;
}