diff --git a/lib/pengine/native.c b/lib/pengine/native.c index d3dd437d57..8f3e6d82ed 100644 --- a/lib/pengine/native.c +++ b/lib/pengine/native.c @@ -1,550 +1,550 @@ /* * Copyright (C) 2004 Andrew Beekhof * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include #include #include #include #include #define VARIANT_NATIVE 1 #include "./variant.h" void native_add_running(resource_t * rsc, node_t * node, pe_working_set_t * data_set) { GListPtr gIter = rsc->running_on; CRM_CHECK(node != NULL, return); for (; gIter != NULL; gIter = gIter->next) { node_t *a_node = (node_t *) gIter->data; CRM_CHECK(a_node != NULL, return); if (safe_str_eq(a_node->details->id, node->details->id)) { return; } } pe_rsc_trace(rsc, "Adding %s to %s", rsc->id, node->details->uname); rsc->running_on = g_list_append(rsc->running_on, node); if (rsc->variant == pe_native) { node->details->running_rsc = g_list_append(node->details->running_rsc, rsc); } if (is_not_set(rsc->flags, pe_rsc_managed)) { pe_rsc_info(rsc, "resource %s isnt managed", rsc->id); resource_location(rsc, node, INFINITY, "not_managed_default", data_set); return; } if (rsc->variant == pe_native && g_list_length(rsc->running_on) > 1) { switch (rsc->recovery_type) { case recovery_stop_only: { GHashTableIter gIter; node_t *local_node = NULL; /* make sure it doesnt come up again */ g_hash_table_destroy(rsc->allowed_nodes); rsc->allowed_nodes = node_hash_from_list(data_set->nodes); g_hash_table_iter_init(&gIter, rsc->allowed_nodes); while (g_hash_table_iter_next(&gIter, NULL, (void **)&local_node)) { local_node->weight = -INFINITY; } } break; case recovery_stop_start: break; case recovery_block: clear_bit(rsc->flags, pe_rsc_managed); set_bit(rsc->flags, pe_rsc_block); break; } crm_debug("%s is active on %d nodes including %s: %s", rsc->id, g_list_length(rsc->running_on), node->details->uname, recovery2text(rsc->recovery_type)); } else { pe_rsc_trace(rsc, "Resource %s is active on: %s", rsc->id, node->details->uname); } if (rsc->parent != NULL) { native_add_running(rsc->parent, node, data_set); } } extern void force_non_unique_clone(resource_t * rsc, const char *rid, pe_working_set_t * data_set); gboolean native_unpack(resource_t * rsc, pe_working_set_t * data_set) { native_variant_data_t *native_data = NULL; pe_rsc_trace(rsc, "Processing resource %s...", rsc->id); native_data = calloc(1, sizeof(native_variant_data_t)); if (is_set(rsc->flags, pe_rsc_unique) && rsc->parent) { const char *class = crm_element_value(rsc->xml, XML_AGENT_ATTR_CLASS); if (safe_str_eq(class, "lsb")) { resource_t *top = uber_parent(rsc); force_non_unique_clone(top, rsc->id, data_set); } } rsc->variant_opaque = native_data; return TRUE; } resource_t * native_find_rsc(resource_t * rsc, const char *id, node_t * on_node, int flags) { gboolean match = FALSE; resource_t *result = NULL; GListPtr gIter = rsc->children; CRM_ASSERT(id != NULL); if (flags & pe_find_clone) { const char *rid = ID(rsc->xml); if(rsc->parent == NULL) { match = FALSE; } else if (safe_str_eq(rsc->id, id)) { match = TRUE; } else if(safe_str_eq(rid, id)) { match = TRUE; } } else { if (strcmp(rsc->id, id) == 0) { match = TRUE; } else if (is_set(flags, pe_find_renamed) && rsc->clone_name && strcmp(rsc->clone_name, id) == 0) { match = TRUE; } } if (match && on_node) { pe_rsc_trace(rsc, "Now checking %s is on %s", rsc->id, on_node->details->uname); if (is_set(flags, pe_find_current) && rsc->running_on) { GListPtr gIter = rsc->running_on; for (; gIter != NULL; gIter = gIter->next) { node_t *loc = (node_t *) gIter->data; if (loc->details == on_node->details) { return rsc; } } } else if (is_set(flags, pe_find_inactive) && rsc->running_on == NULL) { return rsc; } else if (is_not_set(flags, pe_find_current) && rsc->allocated_to && rsc->allocated_to->details == on_node->details) { return rsc; } } else if (match) { return rsc; } for (; gIter != NULL; gIter = gIter->next) { resource_t *child = (resource_t *) gIter->data; result = rsc->fns->find_rsc(child, id, on_node, flags); if (result) { return result; } } return NULL; } char * native_parameter(resource_t * rsc, node_t * node, gboolean create, const char *name, pe_working_set_t * data_set) { char *value_copy = NULL; const char *value = NULL; GHashTable *hash = rsc->parameters; GHashTable *local_hash = NULL; CRM_CHECK(rsc != NULL, return NULL); CRM_CHECK(name != NULL && strlen(name) != 0, return NULL); pe_rsc_trace(rsc, "Looking up %s in %s", name, rsc->id); if (create || g_hash_table_size(rsc->parameters) == 0) { if (node != NULL) { pe_rsc_trace(rsc, "Creating hash with node %s", node->details->uname); } else { pe_rsc_trace(rsc, "Creating default hash"); } local_hash = g_hash_table_new_full(crm_str_hash, g_str_equal, g_hash_destroy_str, g_hash_destroy_str); get_rsc_attributes(local_hash, rsc, node, data_set); hash = local_hash; } value = g_hash_table_lookup(hash, name); if (value == NULL) { /* try meta attributes instead */ value = g_hash_table_lookup(rsc->meta, name); } if (value != NULL) { value_copy = strdup(value); } if (local_hash != NULL) { g_hash_table_destroy(local_hash); } return value_copy; } gboolean native_active(resource_t * rsc, gboolean all) { GListPtr gIter = rsc->running_on; for (; gIter != NULL; gIter = gIter->next) { node_t *a_node = (node_t *) gIter->data; if (a_node->details->unclean) { crm_debug("Resource %s: node %s is unclean", rsc->id, a_node->details->uname); return TRUE; } else if (a_node->details->online == FALSE) { crm_debug("Resource %s: node %s is offline", rsc->id, a_node->details->uname); } else { crm_debug("Resource %s active on %s", rsc->id, a_node->details->uname); return TRUE; } } return FALSE; } struct print_data_s { long options; void *print_data; }; static void native_print_attr(gpointer key, gpointer value, gpointer user_data) { long options = ((struct print_data_s *)user_data)->options; void *print_data = ((struct print_data_s *)user_data)->print_data; status_print("Option: %s = %s\n", (char *)key, (char *)value); } static void native_print_xml(resource_t * rsc, const char *pre_text, long options, void *print_data) { const char *class = crm_element_value(rsc->xml, XML_AGENT_ATTR_CLASS); const char *prov = crm_element_value(rsc->xml, XML_AGENT_ATTR_PROVIDER); /* resource information. */ status_print("%sid); status_print("resource_agent=\"%s%s%s:%s\" ", class, prov ? "::" : "", prov ? prov : "", crm_element_value(rsc->xml, XML_ATTR_TYPE)); status_print("role=\"%s\" ", role2text(rsc->role)); status_print("active=\"%s\" ", rsc->fns->active(rsc, TRUE) ? "true" : "false"); status_print("orphaned=\"%s\" ", is_set(rsc->flags, pe_rsc_orphan) ? "true" : "false"); status_print("managed=\"%s\" ", is_set(rsc->flags, pe_rsc_managed) ? "true" : "false"); status_print("failed=\"%s\" ", is_set(rsc->flags, pe_rsc_failed) ? "true" : "false"); status_print("failure_ignored=\"%s\" ", is_set(rsc->flags, pe_rsc_failure_ignored) ? "true" : "false"); status_print("nodes_running_on=\"%d\" ", g_list_length(rsc->running_on)); if (options & pe_print_dev) { status_print("provisional=\"%s\" ", is_set(rsc->flags, pe_rsc_provisional) ? "true" : "false"); status_print("runnable=\"%s\" ", is_set(rsc->flags, pe_rsc_runnable) ? "true" : "false"); status_print("priority=\"%f\" ", (double)rsc->priority); status_print("variant=\"%s\" ", crm_element_name(rsc->xml)); } /* print out the nodes this resource is running on */ if (options & pe_print_rsconly) { status_print("/>\n"); /* do nothing */ } else if (g_list_length(rsc->running_on) > 0) { GListPtr gIter = rsc->running_on; status_print(">\n"); for (; gIter != NULL; gIter = gIter->next) { node_t *node = (node_t *) gIter->data; status_print("%s \n", pre_text, node->details->uname, node->details->id, node->details->online ? "false" : "true"); } status_print("%s\n", pre_text); } else { status_print("/>\n"); } } void native_print(resource_t * rsc, const char *pre_text, long options, void *print_data) { node_t *node = NULL; const char *prov = NULL; const char *class = crm_element_value(rsc->xml, XML_AGENT_ATTR_CLASS); if (pre_text == NULL && (options & pe_print_printf)) { pre_text = " "; } if (options & pe_print_xml) { native_print_xml(rsc, pre_text, options, print_data); return; } if (safe_str_eq(class, "ocf")) { prov = crm_element_value(rsc->xml, XML_AGENT_ATTR_PROVIDER); } if (rsc->running_on != NULL) { node = rsc->running_on->data; } if (options & pe_print_html) { if (is_not_set(rsc->flags, pe_rsc_managed)) { status_print(""); } else if (is_set(rsc->flags, pe_rsc_failed)) { status_print(""); } else if (rsc->variant == pe_native && g_list_length(rsc->running_on) == 0) { status_print(""); } else if (g_list_length(rsc->running_on) > 1) { status_print(""); } else if (is_set(rsc->flags, pe_rsc_failure_ignored)) { status_print(""); } else { status_print(""); } } if ((options & pe_print_rsconly) || g_list_length(rsc->running_on) > 1) { const char *desc = NULL; desc = crm_element_value(rsc->xml, XML_ATTR_DESC); - status_print("%s%s\t(%s%s%s:%s%s) %s %s%s%s%s%s", + status_print("%s%s\t(%s%s%s:%s%s):\t%s %s%s%s%s%s", pre_text ? pre_text : "", rsc->id, class, prov ? "::" : "", prov ? prov : "", crm_element_value(rsc->xml, XML_ATTR_TYPE), is_set(rsc->flags, pe_rsc_orphan) ? " ORPHANED" : "", (rsc->variant != pe_native) ? "" : role2text(rsc->role), - is_set(rsc->flags, pe_rsc_managed) ? "" : " (unmanaged)", - is_set(rsc->flags, pe_rsc_failed) ? " FAILED" : "", - is_set(rsc->flags, pe_rsc_failure_ignored) ? " (failure ignored)" : "", + is_set(rsc->flags, pe_rsc_managed) ? "" : "(unmanaged) ", + is_set(rsc->flags, pe_rsc_failed) ? "FAILED " : "", + is_set(rsc->flags, pe_rsc_failure_ignored) ? "(failure ignored) " : "", desc ? ": " : "", desc ? desc : ""); } else { status_print("%s%s\t(%s%s%s:%s):\t%s%s %s%s%s%s", pre_text ? pre_text : "", rsc->id, class, prov ? "::" : "", prov ? prov : "", crm_element_value(rsc->xml, XML_ATTR_TYPE), is_set(rsc->flags, pe_rsc_orphan) ? " ORPHANED " : "", (rsc->variant != pe_native) ? "" : role2text(rsc->role), (rsc->variant != pe_native) ? "" : node != NULL ? node->details->uname : "", is_set(rsc->flags, pe_rsc_managed) ? "" : " (unmanaged)", is_set(rsc->flags, pe_rsc_failed) ? " FAILED" : "", is_set(rsc->flags, pe_rsc_failure_ignored) ? " (failure ignored)" : ""); #if CURSES_ENABLED if (options & pe_print_ncurses) { /* coverity[negative_returns] False positive */ move(-1, 0); } #endif } if (options & pe_print_html) { status_print(" "); } if ((options & pe_print_rsconly)) { } else if (g_list_length(rsc->running_on) > 1) { GListPtr gIter = rsc->running_on; int counter = 0; if (options & pe_print_html) { status_print("
    \n"); } else if ((options & pe_print_printf) || (options & pe_print_ncurses)) { status_print("["); } for (; gIter != NULL; gIter = gIter->next) { node_t *node = (node_t *) gIter->data; counter++; if (options & pe_print_html) { status_print("
  • \n%s", node->details->uname); } else if ((options & pe_print_printf) || (options & pe_print_ncurses)) { - status_print("\t%s", node->details->uname); + status_print(" %s", node->details->uname); } else if ((options & pe_print_log)) { status_print("\t%d : %s", counter, node->details->uname); } else { status_print("%s", node->details->uname); } if (options & pe_print_html) { status_print("
  • \n"); } } if (options & pe_print_html) { status_print("
\n"); } else if ((options & pe_print_printf) || (options & pe_print_ncurses)) { status_print(" ]"); } } if (options & pe_print_html) { status_print("
\n"); } else if (options & pe_print_suppres_nl) { /* nothing */ } else if ((options & pe_print_printf) || (options & pe_print_ncurses)) { status_print("\n"); } if (options & pe_print_details) { struct print_data_s pdata; pdata.options = options; pdata.print_data = print_data; g_hash_table_foreach(rsc->parameters, native_print_attr, &pdata); } if (options & pe_print_dev) { GHashTableIter iter; node_t *node = NULL; status_print("%s\t(%s%svariant=%s, priority=%f)", pre_text, is_set(rsc->flags, pe_rsc_provisional) ? "provisional, " : "", is_set(rsc->flags, pe_rsc_runnable) ? "" : "non-startable, ", crm_element_name(rsc->xml), (double)rsc->priority); status_print("%s\tAllowed Nodes", pre_text); g_hash_table_iter_init(&iter, rsc->allowed_nodes); while (g_hash_table_iter_next(&iter, NULL, (void **)&node)) { status_print("%s\t * %s %d", pre_text, node->details->uname, node->weight); } } if (options & pe_print_max_details) { GHashTableIter iter; node_t *node = NULL; status_print("%s\t=== Allowed Nodes\n", pre_text); g_hash_table_iter_init(&iter, rsc->allowed_nodes); while (g_hash_table_iter_next(&iter, NULL, (void **)&node)) { print_node("\t", node, FALSE); } } } void native_free(resource_t * rsc) { pe_rsc_trace(rsc, "Freeing resource action list (not the data)"); common_free(rsc); } enum rsc_role_e native_resource_state(const resource_t * rsc, gboolean current) { enum rsc_role_e role = rsc->next_role; if (current) { role = rsc->role; } pe_rsc_trace(rsc, "%s state: %s", rsc->id, role2text(role)); return role; } node_t * native_location(resource_t * rsc, GListPtr * list, gboolean current) { node_t *one = NULL; GListPtr result = NULL; if (rsc->children) { GListPtr gIter = rsc->children; for (; gIter != NULL; gIter = gIter->next) { resource_t *child = (resource_t *) gIter->data; child->fns->location(child, &result, current); } } else if (current && rsc->running_on) { result = g_list_copy(rsc->running_on); } else if (current == FALSE && rsc->allocated_to) { result = g_list_append(NULL, rsc->allocated_to); } if (result && g_list_length(result) == 1) { one = g_list_nth_data(result, 0); } if (list) { GListPtr gIter = result; for (; gIter != NULL; gIter = gIter->next) { node_t *node = (node_t *) gIter->data; if (*list == NULL || pe_find_node_id(*list, node->details->id) == NULL) { *list = g_list_append(*list, node); } } } g_list_free(result); return one; } diff --git a/pengine/test10/797.summary b/pengine/test10/797.summary index d0509cf266..b7a5da715a 100644 --- a/pengine/test10/797.summary +++ b/pengine/test10/797.summary @@ -1,74 +1,74 @@ Current cluster status: Node c001n08 (6427cb5a-c7a5-4bdf-9892-a04ce56f4e6b): UNCLEAN (offline) Online: [ c001n01 c001n02 c001n03 ] DcIPaddr (ocf::heartbeat:IPaddr): Started c001n03 rsc_c001n08 (ocf::heartbeat:IPaddr): Started c001n02 rsc_c001n02 (ocf::heartbeat:IPaddr): Started c001n02 rsc_c001n03 (ocf::heartbeat:IPaddr): Started c001n03 rsc_c001n01 (ocf::heartbeat:IPaddr): Started c001n01 Clone Set: DoFencing [child_DoFencing] (unique) - child_DoFencing:0 (stonith:ssh) Started [ c001n01 c001n03 ] + child_DoFencing:0 (stonith:ssh): Started [ c001n01 c001n03 ] child_DoFencing:1 (stonith:ssh): Started c001n02 child_DoFencing:2 (stonith:ssh): Started c001n03 child_DoFencing:3 (stonith:ssh): Stopped Transition Summary: * Stop DcIPaddr (Started c001n03) * Move rsc_c001n08 (Started c001n02 -> c001n01) * Move rsc_c001n02 (Started c001n02 -> c001n01) * Stop rsc_c001n03 (Started c001n03) * Stop rsc_c001n01 (Started c001n01) * Restart child_DoFencing:0 (Started c001n01) * Stop child_DoFencing:1 (c001n02) Executing cluster transition: * Resource action: DcIPaddr monitor on c001n02 * Resource action: DcIPaddr monitor on c001n01 * Resource action: rsc_c001n08 stop on c001n02 * Resource action: rsc_c001n08 monitor on c001n03 * Resource action: rsc_c001n08 monitor on c001n01 * Resource action: rsc_c001n02 stop on c001n02 * Resource action: rsc_c001n02 monitor on c001n03 * Resource action: rsc_c001n02 monitor on c001n01 * Resource action: rsc_c001n03 monitor on c001n02 * Resource action: rsc_c001n03 monitor on c001n01 * Resource action: rsc_c001n01 monitor on c001n03 * Resource action: child_DoFencing:2 monitor on c001n01 * Resource action: child_DoFencing:3 monitor on c001n03 * Resource action: child_DoFencing:3 monitor on c001n02 * Resource action: child_DoFencing:3 monitor on c001n01 * Pseudo action: DoFencing_stop_0 * Pseudo action: probe_complete * Resource action: DcIPaddr stop on c001n03 * Resource action: rsc_c001n03 stop on c001n03 * Resource action: rsc_c001n01 stop on c001n01 * Resource action: child_DoFencing:0 stop on c001n03 * Resource action: child_DoFencing:0 stop on c001n01 * Pseudo action: all_stopped * Resource action: DcIPaddr delete on c001n03 * Resource action: child_DoFencing:1 stop on c001n02 * Pseudo action: DoFencing_stopped_0 * Pseudo action: DoFencing_start_0 * Cluster action: do_shutdown on c001n02 * Cluster action: lrm_refresh on c001n03 * Resource action: child_DoFencing:0 start on c001n01 * Resource action: child_DoFencing:0 monitor=5000 on c001n01 * Pseudo action: DoFencing_running_0 Revised cluster status: Node c001n08 (6427cb5a-c7a5-4bdf-9892-a04ce56f4e6b): UNCLEAN (offline) Online: [ c001n01 c001n02 c001n03 ] DcIPaddr (ocf::heartbeat:IPaddr): Started c001n03 rsc_c001n08 (ocf::heartbeat:IPaddr): Stopped rsc_c001n02 (ocf::heartbeat:IPaddr): Stopped rsc_c001n03 (ocf::heartbeat:IPaddr): Stopped rsc_c001n01 (ocf::heartbeat:IPaddr): Stopped Clone Set: DoFencing [child_DoFencing] (unique) child_DoFencing:0 (stonith:ssh): Started c001n01 child_DoFencing:1 (stonith:ssh): Stopped child_DoFencing:2 (stonith:ssh): Started c001n03 child_DoFencing:3 (stonith:ssh): Stopped diff --git a/pengine/test10/inc6.summary b/pengine/test10/inc6.summary index 108de5ba88..ea5f085c33 100644 --- a/pengine/test10/inc6.summary +++ b/pengine/test10/inc6.summary @@ -1,108 +1,108 @@ Current cluster status: Online: [ node1 node2 ] Clone Set: rsc1 [child_rsc1] Started: [ node1 node2 ] Stopped: [ child_rsc1:2 ] Clone Set: rsc2 [child_rsc2] (unique) child_rsc2:0 (heartbeat:apache): Started node1 child_rsc2:1 (heartbeat:apache): Started node1 child_rsc2:2 (heartbeat:apache): Stopped Clone Set: rsc3 [child_rsc3] Started: [ node1 node2 ] Stopped: [ child_rsc3:2 ] Clone Set: rsc4 [child_rsc4] (unique) child_rsc4:0 (heartbeat:apache): Started node1 child_rsc4:1 (heartbeat:apache): Started node1 child_rsc4:2 (heartbeat:apache): Stopped Clone Set: rsc5 [child_rsc5] (unique) child_rsc5:0 (heartbeat:apache): Started node2 child_rsc5:1 (heartbeat:apache): Started node2 child_rsc5:2 (heartbeat:apache): Stopped Clone Set: rsc6 [child_rsc6] Started: [ node1 node2 ] Stopped: [ child_rsc6:2 ] Clone Set: rsc7 [child_rsc7] (unique) child_rsc7:0 (heartbeat:apache): Started node2 child_rsc7:1 (heartbeat:apache): Started node2 child_rsc7:2 (heartbeat:apache): Stopped Clone Set: rsc8 [child_rsc8] Started: [ node1 node2 ] Stopped: [ child_rsc8:2 ] Transition Summary: * Move child_rsc2:1 (Started node1 -> node2) * Move child_rsc4:1 (Started node1 -> node2) * Move child_rsc5:1 (Started node2 -> node1) * Restart child_rsc6:0 (Started node1) * Restart child_rsc6:1 (Started node2) * Move child_rsc7:1 (Started node2 -> node1) Executing cluster transition: * Pseudo action: rsc2_stop_0 * Pseudo action: rsc4_stop_0 * Pseudo action: rsc6_stop_0 * Pseudo action: rsc7_stop_0 * Resource action: child_rsc2:1 stop on node1 * Pseudo action: rsc2_stopped_0 * Pseudo action: rsc2_start_0 * Resource action: child_rsc4:1 stop on node1 * Pseudo action: rsc4_stopped_0 * Pseudo action: rsc4_start_0 * Resource action: child_rsc6:0 stop on node1 * Resource action: child_rsc6:1 stop on node2 * Pseudo action: rsc6_stopped_0 * Resource action: child_rsc7:1 stop on node2 * Pseudo action: rsc7_stopped_0 * Pseudo action: rsc7_start_0 * Resource action: child_rsc2:1 start on node2 * Pseudo action: rsc2_running_0 * Resource action: child_rsc4:1 start on node2 * Pseudo action: rsc4_running_0 * Pseudo action: rsc5_stop_0 * Resource action: child_rsc7:1 start on node1 * Pseudo action: rsc7_running_0 * Resource action: child_rsc5:1 stop on node2 * Pseudo action: rsc5_stopped_0 * Pseudo action: rsc5_start_0 * Pseudo action: all_stopped * Resource action: child_rsc5:1 start on node1 * Pseudo action: rsc5_running_0 * Pseudo action: rsc6_start_0 * Resource action: child_rsc6:0 start on node1 * Resource action: child_rsc6:1 start on node2 * Pseudo action: rsc6_running_0 Revised cluster status: Online: [ node1 node2 ] Clone Set: rsc1 [child_rsc1] Started: [ node1 node2 ] Stopped: [ child_rsc1:2 ] Clone Set: rsc2 [child_rsc2] (unique) child_rsc2:0 (heartbeat:apache): Started node1 - child_rsc2:1 (heartbeat:apache) Started [ node1 node2 ] + child_rsc2:1 (heartbeat:apache): Started [ node1 node2 ] child_rsc2:2 (heartbeat:apache): Stopped Clone Set: rsc3 [child_rsc3] Started: [ node1 node2 ] Stopped: [ child_rsc3:2 ] Clone Set: rsc4 [child_rsc4] (unique) child_rsc4:0 (heartbeat:apache): Started node1 - child_rsc4:1 (heartbeat:apache) Started [ node1 node2 ] + child_rsc4:1 (heartbeat:apache): Started [ node1 node2 ] child_rsc4:2 (heartbeat:apache): Stopped Clone Set: rsc5 [child_rsc5] (unique) child_rsc5:0 (heartbeat:apache): Started node2 child_rsc5:1 (heartbeat:apache): Started node1 child_rsc5:2 (heartbeat:apache): Stopped Clone Set: rsc6 [child_rsc6] Started: [ node1 node2 ] Stopped: [ child_rsc6:2 ] Clone Set: rsc7 [child_rsc7] (unique) child_rsc7:0 (heartbeat:apache): Started node2 child_rsc7:1 (heartbeat:apache): Started node1 child_rsc7:2 (heartbeat:apache): Stopped Clone Set: rsc8 [child_rsc8] Started: [ node1 node2 ] Stopped: [ child_rsc8:2 ] diff --git a/pengine/test10/migrate-fail-2.summary b/pengine/test10/migrate-fail-2.summary index 255a650fa1..1d795d2377 100644 --- a/pengine/test10/migrate-fail-2.summary +++ b/pengine/test10/migrate-fail-2.summary @@ -1,26 +1,26 @@ Current cluster status: Online: [ hex-13 hex-14 ] - test-vm (ocf::heartbeat:Xen) Started FAILED[ hex-13 hex-14 ] + test-vm (ocf::heartbeat:Xen): Started FAILED [ hex-13 hex-14 ] Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] Transition Summary: * Recover test-vm (Started hex-13) Executing cluster transition: * Resource action: test-vm stop on hex-14 * Resource action: test-vm stop on hex-13 * Pseudo action: load_stopped_hex-14 * Pseudo action: load_stopped_hex-13 * Pseudo action: all_stopped * Resource action: test-vm start on hex-13 Revised cluster status: Online: [ hex-13 hex-14 ] test-vm (ocf::heartbeat:Xen): Started hex-13 Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] diff --git a/pengine/test10/migrate-fail-4.summary b/pengine/test10/migrate-fail-4.summary index cf5801d9d7..cda25adede 100644 --- a/pengine/test10/migrate-fail-4.summary +++ b/pengine/test10/migrate-fail-4.summary @@ -1,26 +1,26 @@ Current cluster status: Online: [ hex-13 hex-14 ] - test-vm (ocf::heartbeat:Xen) Started [ hex-13 hex-14 ] + test-vm (ocf::heartbeat:Xen): Started [ hex-13 hex-14 ] Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] Transition Summary: * Restart test-vm (Started hex-13) Executing cluster transition: * Resource action: test-vm stop on hex-14 * Resource action: test-vm stop on hex-13 * Pseudo action: load_stopped_hex-14 * Pseudo action: load_stopped_hex-13 * Pseudo action: all_stopped * Resource action: test-vm start on hex-13 Revised cluster status: Online: [ hex-13 hex-14 ] test-vm (ocf::heartbeat:Xen): Started hex-13 Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] diff --git a/pengine/test10/migrate-fail-6.summary b/pengine/test10/migrate-fail-6.summary index 0433f14b2f..be51c7ac93 100644 --- a/pengine/test10/migrate-fail-6.summary +++ b/pengine/test10/migrate-fail-6.summary @@ -1,26 +1,26 @@ Current cluster status: Online: [ hex-13 hex-14 ] - test-vm (ocf::heartbeat:Xen) Started FAILED[ hex-13 hex-14 ] + test-vm (ocf::heartbeat:Xen): Started FAILED [ hex-13 hex-14 ] Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] Transition Summary: * Recover test-vm (Started hex-13) Executing cluster transition: * Resource action: test-vm stop on hex-13 * Resource action: test-vm stop on hex-14 * Pseudo action: load_stopped_hex-14 * Pseudo action: load_stopped_hex-13 * Pseudo action: all_stopped * Resource action: test-vm start on hex-13 Revised cluster status: Online: [ hex-13 hex-14 ] test-vm (ocf::heartbeat:Xen): Started hex-13 Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] diff --git a/pengine/test10/migrate-partial-2.summary b/pengine/test10/migrate-partial-2.summary index 60fd466cbb..94f84368b3 100644 --- a/pengine/test10/migrate-partial-2.summary +++ b/pengine/test10/migrate-partial-2.summary @@ -1,26 +1,26 @@ Current cluster status: Online: [ hex-13 hex-14 ] - test-vm (ocf::heartbeat:Xen) Started [ hex-13 hex-14 ] + test-vm (ocf::heartbeat:Xen): Started [ hex-13 hex-14 ] Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] Transition Summary: * Migrate test-vm (Started hex-14 -> hex-13) Executing cluster transition: * Pseudo action: load_stopped_hex-13 * Resource action: test-vm migrate_from on hex-13 * Resource action: test-vm stop on hex-14 * Pseudo action: load_stopped_hex-14 * Pseudo action: all_stopped * Pseudo action: test-vm_start_0 Revised cluster status: Online: [ hex-13 hex-14 ] test-vm (ocf::heartbeat:Xen): Started hex-13 Clone Set: c-clusterfs [dlm] Started: [ hex-13 hex-14 ] diff --git a/pengine/test10/mon-rsc-4.summary b/pengine/test10/mon-rsc-4.summary index f671b190f7..18dbccd8aa 100644 --- a/pengine/test10/mon-rsc-4.summary +++ b/pengine/test10/mon-rsc-4.summary @@ -1,24 +1,24 @@ Current cluster status: Node node2 (uuid2): standby Online: [ node1 ] rsc1 (heartbeat:apache): Started node2 Transition Summary: * Move rsc1 (Started node2 -> node1) Executing cluster transition: * Resource action: rsc1 monitor on node1 * Pseudo action: probe_complete * Resource action: rsc1 stop on node2 * Resource action: rsc1 start on node1 * Pseudo action: all_stopped * Resource action: rsc1 monitor=5000 on node1 Revised cluster status: Node node2 (uuid2): standby Online: [ node1 ] - rsc1 (heartbeat:apache) Started [ node1 node2 ] + rsc1 (heartbeat:apache): Started [ node1 node2 ] diff --git a/pengine/test10/multi1.summary b/pengine/test10/multi1.summary index f5d5a10ebe..c782e98efc 100644 --- a/pengine/test10/multi1.summary +++ b/pengine/test10/multi1.summary @@ -1,20 +1,20 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started [ node1 node2 ] + rsc1 (heartbeat:apache): Started [ node1 node2 ] Transition Summary: * Restart rsc1 (Started node1) Executing cluster transition: * Resource action: rsc1 stop on node2 * Resource action: rsc1 stop on node1 * Pseudo action: all_stopped * Resource action: rsc1 start on node1 Revised cluster status: Online: [ node1 node2 ] rsc1 (heartbeat:apache): Started node1 diff --git a/pengine/test10/rec-rsc-0.summary b/pengine/test10/rec-rsc-0.summary index 6c49b376fb..6607ffc702 100644 --- a/pengine/test10/rec-rsc-0.summary +++ b/pengine/test10/rec-rsc-0.summary @@ -1,20 +1,20 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started FAILED[ node1 node2 ] + rsc1 (heartbeat:apache): Started FAILED [ node1 node2 ] Transition Summary: * Stop rsc1 (node1) * Stop rsc1 (node2) Executing cluster transition: * Resource action: rsc1 stop on node2 * Resource action: rsc1 stop on node1 * Pseudo action: all_stopped Revised cluster status: Online: [ node1 node2 ] rsc1 (heartbeat:apache): Stopped diff --git a/pengine/test10/rec-rsc-6.summary b/pengine/test10/rec-rsc-6.summary index f5d5a10ebe..c782e98efc 100644 --- a/pengine/test10/rec-rsc-6.summary +++ b/pengine/test10/rec-rsc-6.summary @@ -1,20 +1,20 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started [ node1 node2 ] + rsc1 (heartbeat:apache): Started [ node1 node2 ] Transition Summary: * Restart rsc1 (Started node1) Executing cluster transition: * Resource action: rsc1 stop on node2 * Resource action: rsc1 stop on node1 * Pseudo action: all_stopped * Resource action: rsc1 start on node1 Revised cluster status: Online: [ node1 node2 ] rsc1 (heartbeat:apache): Started node1 diff --git a/pengine/test10/rec-rsc-7.summary b/pengine/test10/rec-rsc-7.summary index f600367536..d32bdba23c 100644 --- a/pengine/test10/rec-rsc-7.summary +++ b/pengine/test10/rec-rsc-7.summary @@ -1,20 +1,20 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started [ node1 node2 ] + rsc1 (heartbeat:apache): Started [ node1 node2 ] Transition Summary: * Stop rsc1 (node1) * Stop rsc1 (node2) Executing cluster transition: * Resource action: rsc1 stop on node2 * Resource action: rsc1 stop on node1 * Pseudo action: all_stopped Revised cluster status: Online: [ node1 node2 ] rsc1 (heartbeat:apache): Stopped diff --git a/pengine/test10/rec-rsc-8.summary b/pengine/test10/rec-rsc-8.summary index 1ab08bb364..35510e8f18 100644 --- a/pengine/test10/rec-rsc-8.summary +++ b/pengine/test10/rec-rsc-8.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started (unmanaged)[ node1 node2 ] + rsc1 (heartbeat:apache): Started (unmanaged) [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (heartbeat:apache) Started (unmanaged)[ node1 node2 ] + rsc1 (heartbeat:apache): Started (unmanaged) [ node1 node2 ] diff --git a/pengine/test10/stonith-0.summary b/pengine/test10/stonith-0.summary index 427adff4ac..2ff08c34aa 100644 --- a/pengine/test10/stonith-0.summary +++ b/pengine/test10/stonith-0.summary @@ -1,110 +1,110 @@ Current cluster status: Node c001n03 (f5e1d2de-73da-432a-9d5c-37472253c2ee): UNCLEAN (online) Node c001n05 (52a5ea5e-86ee-442c-b251-0bc9825c517e): UNCLEAN (online) Online: [ c001n02 c001n04 c001n06 c001n07 c001n08 ] DcIPaddr (ocf::heartbeat:IPaddr): Stopped Resource Group: group-1 - ocf_192.168.100.181 (ocf::heartbeat:IPaddr) Started [ c001n03 c001n05 ] + ocf_192.168.100.181 (ocf::heartbeat:IPaddr): Started [ c001n03 c001n05 ] heartbeat_192.168.100.182 (heartbeat:IPaddr): Started c001n03 - ocf_192.168.100.183 (ocf::heartbeat:IPaddr) Started FAILED[ c001n03 c001n05 ] + ocf_192.168.100.183 (ocf::heartbeat:IPaddr): Started FAILED [ c001n03 c001n05 ] lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n04 rsc_c001n03 (ocf::heartbeat:IPaddr): Started c001n06 rsc_c001n02 (ocf::heartbeat:IPaddr): Started c001n02 rsc_c001n04 (ocf::heartbeat:IPaddr): Started c001n04 rsc_c001n05 (ocf::heartbeat:IPaddr): Started c001n05 rsc_c001n06 (ocf::heartbeat:IPaddr): Started c001n06 rsc_c001n07 (ocf::heartbeat:IPaddr): Started c001n03 rsc_c001n08 (ocf::heartbeat:IPaddr): Started c001n08 Clone Set: DoFencing [child_DoFencing] Started: [ c001n02 c001n04 c001n06 c001n07 c001n08 ] Stopped: [ child_DoFencing:5 child_DoFencing:6 ] Master/Slave Set: master_rsc_1 [ocf_msdummy] (unique) ocf_msdummy:0 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Master c001n02 ocf_msdummy:1 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n02 ocf_msdummy:2 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n07 ocf_msdummy:3 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n07 ocf_msdummy:4 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Slave c001n08 ocf_msdummy:5 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Slave c001n08 ocf_msdummy:6 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:7 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:8 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:9 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:10 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n04 ocf_msdummy:11 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n04 ocf_msdummy:12 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n06 ocf_msdummy:13 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n06 Transition Summary: * Move ocf_192.168.100.181 (Started c001n03 -> c001n02) * Move heartbeat_192.168.100.182 (Started c001n03 -> c001n02) * Move ocf_192.168.100.183 (Started c001n03 -> c001n02) * Move rsc_c001n05 (Started c001n05 -> c001n07) * Move rsc_c001n07 (Started c001n03 -> c001n07) Executing cluster transition: * Resource action: child_DoFencing:4 monitor=20000 on c001n08 * Pseudo action: stonith_up * Fencing c001n05 * Fencing c001n03 * Pseudo action: stonith_complete * Pseudo action: group-1_stop_0 * Pseudo action: ocf_192.168.100.183_stop_0 * Pseudo action: ocf_192.168.100.183_stop_0 * Pseudo action: rsc_c001n05_stop_0 * Pseudo action: rsc_c001n07_stop_0 * Pseudo action: heartbeat_192.168.100.182_stop_0 * Resource action: rsc_c001n05 start on c001n07 * Resource action: rsc_c001n07 start on c001n07 * Pseudo action: ocf_192.168.100.181_stop_0 * Pseudo action: ocf_192.168.100.181_stop_0 * Resource action: rsc_c001n05 monitor=5000 on c001n07 * Resource action: rsc_c001n07 monitor=5000 on c001n07 * Pseudo action: all_stopped * Pseudo action: group-1_stopped_0 * Pseudo action: group-1_start_0 * Resource action: ocf_192.168.100.181 start on c001n02 * Resource action: heartbeat_192.168.100.182 start on c001n02 * Resource action: ocf_192.168.100.183 start on c001n02 * Pseudo action: group-1_running_0 * Resource action: ocf_192.168.100.181 monitor=5000 on c001n02 * Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02 * Resource action: ocf_192.168.100.183 monitor=5000 on c001n02 Revised cluster status: Online: [ c001n02 c001n04 c001n06 c001n07 c001n08 ] OFFLINE: [ c001n03 c001n05 ] DcIPaddr (ocf::heartbeat:IPaddr): Stopped Resource Group: group-1 ocf_192.168.100.181 (ocf::heartbeat:IPaddr): Started c001n02 heartbeat_192.168.100.182 (heartbeat:IPaddr): Started c001n02 ocf_192.168.100.183 (ocf::heartbeat:IPaddr): Started c001n02 lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n04 rsc_c001n03 (ocf::heartbeat:IPaddr): Started c001n06 rsc_c001n02 (ocf::heartbeat:IPaddr): Started c001n02 rsc_c001n04 (ocf::heartbeat:IPaddr): Started c001n04 rsc_c001n05 (ocf::heartbeat:IPaddr): Started c001n07 rsc_c001n06 (ocf::heartbeat:IPaddr): Started c001n06 rsc_c001n07 (ocf::heartbeat:IPaddr): Started c001n07 rsc_c001n08 (ocf::heartbeat:IPaddr): Started c001n08 Clone Set: DoFencing [child_DoFencing] Started: [ c001n02 c001n04 c001n06 c001n07 c001n08 ] Stopped: [ child_DoFencing:5 child_DoFencing:6 ] Master/Slave Set: master_rsc_1 [ocf_msdummy] (unique) ocf_msdummy:0 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Master c001n02 ocf_msdummy:1 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n02 ocf_msdummy:2 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n07 ocf_msdummy:3 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n07 ocf_msdummy:4 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Slave c001n08 ocf_msdummy:5 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Slave c001n08 ocf_msdummy:6 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:7 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:8 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:9 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped ocf_msdummy:10 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n04 ocf_msdummy:11 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n04 ocf_msdummy:12 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n06 ocf_msdummy:13 (ocf::heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Started c001n06 diff --git a/pengine/test10/stopped-monitor-02.summary b/pengine/test10/stopped-monitor-02.summary index 5728cf525b..03a9517faa 100644 --- a/pengine/test10/stopped-monitor-02.summary +++ b/pengine/test10/stopped-monitor-02.summary @@ -1,22 +1,22 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started FAILED [ node1 node2 ] Transition Summary: * Recover rsc1 (Started node1) Executing cluster transition: * Resource action: rsc1 stop on node1 * Resource action: rsc1 stop on node2 * Resource action: rsc1 monitor=20000 on node2 * Pseudo action: all_stopped * Resource action: rsc1 start on node1 * Resource action: rsc1 monitor=10000 on node1 Revised cluster status: Online: [ node1 node2 ] rsc1 (ocf::pacemaker:Dummy): Started node1 diff --git a/pengine/test10/stopped-monitor-06.summary b/pengine/test10/stopped-monitor-06.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-06.summary +++ b/pengine/test10/stopped-monitor-06.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/pengine/test10/stopped-monitor-07.summary b/pengine/test10/stopped-monitor-07.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-07.summary +++ b/pengine/test10/stopped-monitor-07.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/pengine/test10/stopped-monitor-10.summary b/pengine/test10/stopped-monitor-10.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-10.summary +++ b/pengine/test10/stopped-monitor-10.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/pengine/test10/stopped-monitor-12.summary b/pengine/test10/stopped-monitor-12.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-12.summary +++ b/pengine/test10/stopped-monitor-12.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/pengine/test10/stopped-monitor-22.summary b/pengine/test10/stopped-monitor-22.summary index 1e0935b781..489f7308b5 100644 --- a/pengine/test10/stopped-monitor-22.summary +++ b/pengine/test10/stopped-monitor-22.summary @@ -1,22 +1,22 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started FAILED [ node1 node2 ] Transition Summary: * Stop rsc1 (node1) * Stop rsc1 (node2) Executing cluster transition: * Resource action: rsc1 stop on node2 * Resource action: rsc1 monitor=20000 on node2 * Resource action: rsc1 stop on node1 * Resource action: rsc1 monitor=20000 on node1 * Pseudo action: all_stopped Revised cluster status: Online: [ node1 node2 ] rsc1 (ocf::pacemaker:Dummy): Stopped diff --git a/pengine/test10/stopped-monitor-25.summary b/pengine/test10/stopped-monitor-25.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-25.summary +++ b/pengine/test10/stopped-monitor-25.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/pengine/test10/stopped-monitor-27.summary b/pengine/test10/stopped-monitor-27.summary index 78278ffbce..23dc001ddc 100644 --- a/pengine/test10/stopped-monitor-27.summary +++ b/pengine/test10/stopped-monitor-27.summary @@ -1,15 +1,15 @@ Current cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] Transition Summary: Executing cluster transition: Revised cluster status: Online: [ node1 node2 ] - rsc1 (ocf::pacemaker:Dummy) Started (unmanaged) FAILED[ node1 node2 ] + rsc1 (ocf::pacemaker:Dummy): Started (unmanaged) FAILED [ node1 node2 ] diff --git a/tools/regression.exp b/tools/regression.exp index e6d35a67f3..1e9bd442f3 100755 --- a/tools/regression.exp +++ b/tools/regression.exp @@ -1,1228 +1,1228 @@ Date: 2006-01-08 00:00:00Z Date: 2006-W01-7 00:00:00Z * Passed: iso8601 - 2006-W01-7 Date: 2006-01-02 00:00:00Z Date: 2006-W01-1 00:00:00Z * Passed: iso8601 - 2006-W01-1 Date: 2007-01-07 00:00:00Z Date: 2007-W01-7 00:00:00Z * Passed: iso8601 - 2007-W01-7 Date: 2007-01-01 00:00:00Z Date: 2007-W01-1 00:00:00Z * Passed: iso8601 - 2007-W01-1 Date: 2008-01-06 00:00:00Z Date: 2008-W01-7 00:00:00Z * Passed: iso8601 - 2008-W01-7 Date: 2007-12-31 00:00:00Z Date: 2008-W01-1 00:00:00Z * Passed: iso8601 - 2008-W01-1 Date: 2009-01-04 00:00:00Z Date: 2009-W01-7 00:00:00Z * Passed: iso8601 - 2009-W01-7 Date: 2008-12-29 00:00:00Z Date: 2009-W01-1 00:00:00Z * Passed: iso8601 - 2009-W01-1 Date: 2010-01-10 00:00:00Z Date: 2010-W01-7 00:00:00Z * Passed: iso8601 - 2010-W01-7 Date: 2010-01-04 00:00:00Z Date: 2010-W01-1 00:00:00Z * Passed: iso8601 - 2010-W01-1 Date: 2011-01-09 00:00:00Z Date: 2011-W01-7 00:00:00Z * Passed: iso8601 - 2011-W01-7 Date: 2011-01-03 00:00:00Z Date: 2011-W01-1 00:00:00Z * Passed: iso8601 - 2011-W01-1 Date: 2012-01-08 00:00:00Z Date: 2012-W01-7 00:00:00Z * Passed: iso8601 - 2012-W01-7 Date: 2012-01-02 00:00:00Z Date: 2012-W01-1 00:00:00Z * Passed: iso8601 - 2012-W01-1 Date: 2013-01-06 00:00:00Z Date: 2013-W01-7 00:00:00Z * Passed: iso8601 - 2013-W01-7 Date: 2012-12-31 00:00:00Z Date: 2013-W01-1 00:00:00Z * Passed: iso8601 - 2013-W01-1 Date: 2014-01-05 00:00:00Z Date: 2014-W01-7 00:00:00Z * Passed: iso8601 - 2014-W01-7 Date: 2013-12-30 00:00:00Z Date: 2014-W01-1 00:00:00Z * Passed: iso8601 - 2014-W01-1 Date: 2015-01-04 00:00:00Z Date: 2015-W01-7 00:00:00Z * Passed: iso8601 - 2015-W01-7 Date: 2014-12-29 00:00:00Z Date: 2015-W01-1 00:00:00Z * Passed: iso8601 - 2015-W01-1 Date: 2016-01-10 00:00:00Z Date: 2016-W01-7 00:00:00Z * Passed: iso8601 - 2016-W01-7 Date: 2016-01-04 00:00:00Z Date: 2016-W01-1 00:00:00Z * Passed: iso8601 - 2016-W01-1 Date: 2017-01-08 00:00:00Z Date: 2017-W01-7 00:00:00Z * Passed: iso8601 - 2017-W01-7 Date: 2017-01-02 00:00:00Z Date: 2017-W01-1 00:00:00Z * Passed: iso8601 - 2017-W01-1 Date: 2018-01-07 00:00:00Z Date: 2018-W01-7 00:00:00Z * Passed: iso8601 - 2018-W01-7 Date: 2018-01-01 00:00:00Z Date: 2018-W01-1 00:00:00Z * Passed: iso8601 - 2018-W01-1 Date: 2009-W53-7 00:00:00Z * Passed: iso8601 - 2009-W53-07 Date: 2009-01-31 00:00:00Z Duration: 0000-01-00 00:00:00Z Duration ends at: 2009-02-28 00:00:00Z * Passed: iso8601 - 2009-01-31 + 1 Month Date: 2009-01-31 00:00:00Z Duration: 0000-02-00 00:00:00Z Duration ends at: 2009-03-31 00:00:00Z * Passed: iso8601 - 2009-01-31 + 2 Months Date: 2009-01-31 00:00:00Z Duration: 0000-03-00 00:00:00Z Duration ends at: 2009-04-30 00:00:00Z * Passed: iso8601 - 2009-01-31 + 3 Months Date: 2009-03-31 00:00:00Z Duration: 0000--01-00 00:00:00Z Duration ends at: 2009-02-28 00:00:00Z * Passed: iso8601 - 2009-03-31 - 1 Month Setting up shadow instance A new shadow instance was created. To begin using it paste the following into your shell: CIB_shadow=tools-regression ; export CIB_shadow The supplied command is considered dangerous. To prevent accidental destruction of the cluster, the --force flag is required in order to proceed. * Passed: cibadmin - Require --force for CIB erasure * Passed: cibadmin - Allow CIB erasure with --force * Passed: cibadmin - Query CIB * Passed: crm_attribute - Set cluster option * Passed: cibadmin - Query new cluster option * Passed: cibadmin - Query cluster options * Passed: cibadmin - Delete nvpair Call failed: Name not unique on network * Passed: cibadmin - Create operaton should fail with: -76, The object already exists * Passed: cibadmin - Modify cluster options section * Passed: cibadmin - Query updated cluster option * Passed: crm_attribute - Set duplicate cluster option Please choose from one of the matches above and suppy the 'id' with --attr-id * Passed: crm_attribute - Setting multiply defined cluster option should fail with -216, Could not set cluster option * Passed: crm_attribute - Set cluster option with -s Deleted crm_config option: id=(null) name=cluster-delay * Passed: crm_attribute - Delete cluster option with -i * Passed: cibadmin - Create node entry * Passed: cibadmin - Create node status entry * Passed: crm_attribute - Create node attribute * Passed: cibadmin - Query new node attribute Digest: * Passed: cibadmin - Digest calculation Call failed: Update was older than existing configuration * Passed: cibadmin - Replace operation should fail with: -45, Update was older than existing configuration Error performing operation: No such device or address scope=status name=standby value=off * Passed: crm_standby - Default standby value * Passed: crm_standby - Set standby status scope=nodes name=standby value=true * Passed: crm_standby - Query standby value Deleted nodes attribute: id=nodes-clusterNode-UUID-standby name=standby Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) * Passed: crm_standby - Delete standby value * Passed: cibadmin - Create a resource * Passed: crm_resource - Create a resource meta attribute false * Passed: crm_resource - Query a resource meta attribute Deleted dummy option: id=dummy-meta_attributes-is-managed name=is-managed * Passed: crm_resource - Remove a resource meta attribute * Passed: crm_resource - Create a resource attribute - dummy (ocf::pacemaker:Dummy) Stopped + dummy (ocf::pacemaker:Dummy): Stopped * Passed: crm_resource - List the configured resources Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) Could not establish attrd connection: Connection refused (111) * Passed: crm_resource - Set a resource's fail-count Resource dummy not moved: not-active and no preferred location specified. Error performing operation: Invalid argument * Passed: crm_resource - Require a destination when migrating a resource that is stopped Error performing operation: i.dont.exist is not a known node Error performing operation: No such device or address * Passed: crm_resource - Don't support migration to non-existant locations * Passed: crm_resource - Migrate a resource * Passed: crm_resource - Un-migrate a resource false * Passed: crm_ticket - Default ticket granted state * Passed: crm_ticket - Set ticket granted state false * Passed: crm_ticket - Query ticket granted state Deleted ticketA state attribute: name=granted * Passed: crm_ticket - Delete ticket granted state * Passed: crm_ticket - Make a ticket standby true * Passed: crm_ticket - Query ticket standby state * Passed: crm_ticket - Activate a ticket Deleted ticketA state attribute: name=standby * Passed: crm_ticket - Delete ticket standby state