From a7e1531f8b5c6c8d849876097e53f5657fdf9907 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Mon, 18 Nov 2024 17:04:57 -0500 Subject: [PATCH 01/11] API: scheduler: Add pcmk_unpack_scheduler_input. --- include/crm/pengine/status.h | 2 ++ lib/pengine/status.c | 23 ++++++++++++++++------- 2 files changed, 18 insertions(+), 7 deletions(-) diff --git a/include/crm/pengine/status.h b/include/crm/pengine/status.h index d04681d5167..1ccf74aaba1 100644 --- a/include/crm/pengine/status.h +++ b/include/crm/pengine/status.h @@ -30,6 +30,8 @@ extern "C" { const char *rsc_printable_id(const pcmk_resource_t *rsc); +int pcmk_unpack_scheduler_input(pcmk_scheduler_t *scheduler); + // NOTE: sbd (as of at least 1.5.2) uses this gboolean cluster_status(pcmk_scheduler_t *scheduler); diff --git a/lib/pengine/status.c b/lib/pengine/status.c index 1fce0b32af4..d60fa4b11ad 100644 --- a/lib/pengine/status.c +++ b/lib/pengine/status.c @@ -88,8 +88,9 @@ check_for_deprecated_rules(pcmk_scheduler_t *scheduler) } } -/* - * Unpack everything +/*! + * Unpack scheduler input + * * At the end you'll have: * - A list of nodes * - A list of resources (each with any dependencies on other resources) @@ -98,15 +99,17 @@ check_for_deprecated_rules(pcmk_scheduler_t *scheduler) * - A list of nodes that need to be stonith'd * - A list of nodes that need to be shutdown * - A list of the possible stop/start actions (without dependencies) + * + * \return Standard Pacemaker return code */ -gboolean -cluster_status(pcmk_scheduler_t * scheduler) +int +pcmk_unpack_scheduler_input(pcmk_scheduler_t *scheduler) { const char *new_version = NULL; xmlNode *section = NULL; if ((scheduler == NULL) || (scheduler->input == NULL)) { - return FALSE; + return EINVAL; } new_version = crm_element_value(scheduler->input, PCMK_XA_CRM_FEATURE_SET); @@ -114,7 +117,7 @@ cluster_status(pcmk_scheduler_t * scheduler) if (pcmk__check_feature_set(new_version) != pcmk_rc_ok) { pcmk__config_err("Can't process CIB with feature set '%s' greater than our own '%s'", new_version, CRM_FEATURE_SET); - return FALSE; + return pcmk_rc_schema_validation; } crm_trace("Beginning unpack"); @@ -202,7 +205,13 @@ cluster_status(pcmk_scheduler_t * scheduler) } pcmk__set_scheduler_flags(scheduler, pcmk__sched_have_status); - return TRUE; + return pcmk_rc_ok; +} + +gboolean +cluster_status(pcmk_scheduler_t * scheduler) +{ + return pcmk_unpack_scheduler_input(scheduler) == pcmk_rc_ok; } /*! From c0cc157f2a4dfc78fbc830d4a58c47c776f5902d Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Tue, 19 Nov 2024 10:23:41 -0500 Subject: [PATCH 02/11] API: scheduler: Deprecate cluster_status. Use pcmk_unpack_scheduler_input instead. --- daemons/execd/cts-exec-helper.c | 2 +- include/crm/pengine/status.h | 3 --- include/crm/pengine/status_compat.h | 3 +++ lib/pacemaker/pcmk_scheduler.c | 4 ++-- lib/pacemaker/pcmk_simulate.c | 6 +++--- lib/pacemaker/pcmk_status.c | 2 +- lib/pengine/status.c | 12 ++++++------ lib/pengine/tests/native/native_find_rsc_test.c | 2 +- lib/pengine/tests/native/pe_base_name_eq_test.c | 2 +- tools/crm_resource.c | 4 ++-- tools/crm_resource_runtime.c | 2 +- tools/crm_ticket.c | 2 +- 12 files changed, 22 insertions(+), 22 deletions(-) diff --git a/daemons/execd/cts-exec-helper.c b/daemons/execd/cts-exec-helper.c index a74f5993a6c..1c4c1393e13 100644 --- a/daemons/execd/cts-exec-helper.c +++ b/daemons/execd/cts-exec-helper.c @@ -475,7 +475,7 @@ generate_params(void) pcmk__set_scheduler_flags(scheduler, pcmk__sched_no_counts); scheduler->input = cib_xml_copy; scheduler->priv->now = crm_time_new(NULL); - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); // Find resource in CIB rsc = pe_find_resource_with_flags(scheduler->priv->resources, diff --git a/include/crm/pengine/status.h b/include/crm/pengine/status.h index 1ccf74aaba1..bddd8e07d34 100644 --- a/include/crm/pengine/status.h +++ b/include/crm/pengine/status.h @@ -32,9 +32,6 @@ const char *rsc_printable_id(const pcmk_resource_t *rsc); int pcmk_unpack_scheduler_input(pcmk_scheduler_t *scheduler); -// NOTE: sbd (as of at least 1.5.2) uses this -gboolean cluster_status(pcmk_scheduler_t *scheduler); - // NOTE: sbd (as of at least 1.5.2) uses this pcmk_scheduler_t *pe_new_working_set(void); diff --git a/include/crm/pengine/status_compat.h b/include/crm/pengine/status_compat.h index 5488bb8e7d4..e360dcaf41e 100644 --- a/include/crm/pengine/status_compat.h +++ b/include/crm/pengine/status_compat.h @@ -27,6 +27,9 @@ extern "C" { * release. */ +// NOTE: sbd (as of at least 1.5.2) uses this +gboolean cluster_status(pcmk_scheduler_t *scheduler); + // NOTE: sbd (as of at least 1.5.2) uses this //! \deprecated Use pcmk_find_node() with scheduler object instead pcmk_node_t *pe_find_node(const GList *node_list, const char *node_name); diff --git a/lib/pacemaker/pcmk_scheduler.c b/lib/pacemaker/pcmk_scheduler.c index 8960c298f57..24ed01ce288 100644 --- a/lib/pacemaker/pcmk_scheduler.c +++ b/lib/pacemaker/pcmk_scheduler.c @@ -756,7 +756,7 @@ unpack_cib(xmlNode *cib, unsigned long long flags, pcmk_scheduler_t *scheduler) pcmk__set_scheduler_flags(scheduler, flags); scheduler->input = cib; - cluster_status(scheduler); // Sets pcmk__sched_have_status + pcmk_unpack_scheduler_input(scheduler); // Sets pcmk__sched_have_status } /*! @@ -873,7 +873,7 @@ pcmk__init_scheduler(pcmk__output_t *out, xmlNodePtr input, const crm_time_t *da } // Unpack everything - cluster_status(new_scheduler); + pcmk_unpack_scheduler_input(new_scheduler); *scheduler = new_scheduler; return pcmk_rc_ok; diff --git a/lib/pacemaker/pcmk_simulate.c b/lib/pacemaker/pcmk_simulate.c index 7b3080d57e5..9608b32188b 100644 --- a/lib/pacemaker/pcmk_simulate.c +++ b/lib/pacemaker/pcmk_simulate.c @@ -809,7 +809,7 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, } reset(scheduler, input, out, use_date, flags); - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); if (!out->is_quiet(out)) { const bool show_pending = pcmk_is_set(flags, pcmk_sim_show_pending); @@ -862,7 +862,7 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, cleanup_calculations(scheduler); reset(scheduler, input, out, use_date, flags); - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); } if (input_file != NULL) { @@ -972,7 +972,7 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, pcmk__set_scheduler_flags(scheduler, pcmk__sched_show_utilization); } - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); print_cluster_status(scheduler, 0, section_opts, "Revised Cluster Status", true); diff --git a/lib/pacemaker/pcmk_status.c b/lib/pacemaker/pcmk_status.c index 002817c06ae..3e84510cbe4 100644 --- a/lib/pacemaker/pcmk_status.c +++ b/lib/pacemaker/pcmk_status.c @@ -104,7 +104,7 @@ pcmk__output_cluster_status(pcmk_scheduler_t *scheduler, stonith_t *stonith, pe_reset_working_set(scheduler); scheduler->input = cib_copy; - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); /* Unpack constraints if any section will need them * (tickets may be referenced in constraints but not granted yet, diff --git a/lib/pengine/status.c b/lib/pengine/status.c index d60fa4b11ad..d087cfa6fb0 100644 --- a/lib/pengine/status.c +++ b/lib/pengine/status.c @@ -208,12 +208,6 @@ pcmk_unpack_scheduler_input(pcmk_scheduler_t *scheduler) return pcmk_rc_ok; } -gboolean -cluster_status(pcmk_scheduler_t * scheduler) -{ - return pcmk_unpack_scheduler_input(scheduler) == pcmk_rc_ok; -} - /*! * \internal * \brief Free a list of pcmk_resource_t @@ -537,6 +531,12 @@ pe_find_node_id(const GList *nodes, const char *id) #include +gboolean +cluster_status(pcmk_scheduler_t * scheduler) +{ + return pcmk_unpack_scheduler_input(scheduler) == pcmk_rc_ok; +} + /*! * \brief Find a node by name in a list of nodes * diff --git a/lib/pengine/tests/native/native_find_rsc_test.c b/lib/pengine/tests/native/native_find_rsc_test.c index 676d22a88d4..fa8e7e13562 100644 --- a/lib/pengine/tests/native/native_find_rsc_test.c +++ b/lib/pengine/tests/native/native_find_rsc_test.c @@ -46,7 +46,7 @@ setup(void **state) { pcmk__set_scheduler_flags(scheduler, pcmk__sched_no_counts); scheduler->input = input; - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); /* Get references to the cluster nodes so we don't have to find them repeatedly. */ cluster01 = pcmk_find_node(scheduler, "cluster01"); diff --git a/lib/pengine/tests/native/pe_base_name_eq_test.c b/lib/pengine/tests/native/pe_base_name_eq_test.c index 1a08480974b..d8e4eff11af 100644 --- a/lib/pengine/tests/native/pe_base_name_eq_test.c +++ b/lib/pengine/tests/native/pe_base_name_eq_test.c @@ -45,7 +45,7 @@ setup(void **state) { pcmk__set_scheduler_flags(scheduler, pcmk__sched_no_counts); scheduler->input = input; - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); /* Get references to several resources we use frequently. */ for (GList *iter = scheduler->priv->resources; diff --git a/tools/crm_resource.c b/tools/crm_resource.c index b308d847ada..bf9933cd09d 100644 --- a/tools/crm_resource.c +++ b/tools/crm_resource.c @@ -959,7 +959,7 @@ clear_constraints(pcmk__output_t *out) } scheduler->input = cib_xml; - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); after = build_constraint_list(scheduler->input); remaining = pcmk__subtract_lists(before, after, (GCompareFunc) strcmp); @@ -993,7 +993,7 @@ initialize_scheduler_data(xmlNode **cib_xml_orig) return rc; } - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); return pcmk_rc_ok; } diff --git a/tools/crm_resource_runtime.c b/tools/crm_resource_runtime.c index 0c1384ddf48..b2c220fd872 100644 --- a/tools/crm_resource_runtime.c +++ b/tools/crm_resource_runtime.c @@ -1491,7 +1491,7 @@ update_dataset(cib_t *cib, pcmk_scheduler_t *scheduler, xmlNode **cib_xml_orig, pcmk__xml_free(*cib_xml_orig); *cib_xml_orig = xml; - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); } done: diff --git a/tools/crm_ticket.c b/tools/crm_ticket.c index b6644a1779e..7cb3701c1ef 100644 --- a/tools/crm_ticket.c +++ b/tools/crm_ticket.c @@ -441,7 +441,7 @@ main(int argc, char **argv) scheduler->input = cib_xml_copy; scheduler->priv->now = crm_time_new(NULL); - cluster_status(scheduler); + pcmk_unpack_scheduler_input(scheduler); /* For recording the tickets that are referenced in PCMK_XE_RSC_TICKET * constraints but have never been granted yet. From 399de3f1b309b52cbe7fe8e7afcab05280bde825 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Tue, 19 Nov 2024 10:33:12 -0500 Subject: [PATCH 03/11] Refactor: libpacemaker: Don't continue if pcmk_unpack_scheduler_input fails. This function can return a couple error codes, most notably when called on input with a feature set that is newer than the latest supported. In that case, the caller should return its own error instead f trying to continue on with an unpopulated scheduler object. This prevents a cascade of error messages. --- lib/pacemaker/pcmk_scheduler.c | 11 +++++++++-- lib/pacemaker/pcmk_simulate.c | 19 +++++++++++++++---- lib/pacemaker/pcmk_status.c | 9 ++++++++- 3 files changed, 32 insertions(+), 7 deletions(-) diff --git a/lib/pacemaker/pcmk_scheduler.c b/lib/pacemaker/pcmk_scheduler.c index 24ed01ce288..b720749b31a 100644 --- a/lib/pacemaker/pcmk_scheduler.c +++ b/lib/pacemaker/pcmk_scheduler.c @@ -835,6 +835,8 @@ int pcmk__init_scheduler(pcmk__output_t *out, xmlNodePtr input, const crm_time_t *date, pcmk_scheduler_t **scheduler) { + int rc = pcmk_rc_ok; + // Allows for cleaner syntax than dereferencing the scheduler argument pcmk_scheduler_t *new_scheduler = NULL; @@ -857,7 +859,7 @@ pcmk__init_scheduler(pcmk__output_t *out, xmlNodePtr input, const crm_time_t *da } } else { - int rc = cib__signon_query(out, NULL, &(new_scheduler->input)); + rc = cib__signon_query(out, NULL, &(new_scheduler->input)); if (rc != pcmk_rc_ok) { pe_free_working_set(new_scheduler); @@ -873,7 +875,12 @@ pcmk__init_scheduler(pcmk__output_t *out, xmlNodePtr input, const crm_time_t *da } // Unpack everything - pcmk_unpack_scheduler_input(new_scheduler); + rc = pcmk_unpack_scheduler_input(new_scheduler); + if (rc != pcmk_rc_ok) { + pe_free_working_set(new_scheduler); + return rc; + } + *scheduler = new_scheduler; return pcmk_rc_ok; diff --git a/lib/pacemaker/pcmk_simulate.c b/lib/pacemaker/pcmk_simulate.c index 9608b32188b..35c4d215676 100644 --- a/lib/pacemaker/pcmk_simulate.c +++ b/lib/pacemaker/pcmk_simulate.c @@ -809,7 +809,11 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, } reset(scheduler, input, out, use_date, flags); - pcmk_unpack_scheduler_input(scheduler); + rc = pcmk_unpack_scheduler_input(scheduler); + + if (rc != pcmk_rc_ok) { + goto simulate_done; + } if (!out->is_quiet(out)) { const bool show_pending = pcmk_is_set(flags, pcmk_sim_show_pending); @@ -862,6 +866,11 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, cleanup_calculations(scheduler); reset(scheduler, input, out, use_date, flags); + /* pcmk_unpack_scheduler_input only returns error on scheduler being + * NULL or the feature set being unsupported. Neither of those + * conditions could have changed since the first call, so there's no + * need to check the return value again. + */ pcmk_unpack_scheduler_input(scheduler); } @@ -972,9 +981,11 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, pcmk__set_scheduler_flags(scheduler, pcmk__sched_show_utilization); } - pcmk_unpack_scheduler_input(scheduler); - print_cluster_status(scheduler, 0, section_opts, "Revised Cluster Status", - true); + rc = pcmk_unpack_scheduler_input(scheduler); + if (rc == pcmk_rc_ok) { + print_cluster_status(scheduler, 0, section_opts, "Revised Cluster Status", + true); + } simulate_done: cib__clean_up_connection(&cib); diff --git a/lib/pacemaker/pcmk_status.c b/lib/pacemaker/pcmk_status.c index 3e84510cbe4..70f43b03123 100644 --- a/lib/pacemaker/pcmk_status.c +++ b/lib/pacemaker/pcmk_status.c @@ -104,7 +104,14 @@ pcmk__output_cluster_status(pcmk_scheduler_t *scheduler, stonith_t *stonith, pe_reset_working_set(scheduler); scheduler->input = cib_copy; - pcmk_unpack_scheduler_input(scheduler); + rc = pcmk_unpack_scheduler_input(scheduler); + + if (rc != pcmk_rc_ok) { + /* Now that we've set up the scheduler, it's up to the caller to clean up. + * Doing cleanup here can result in double frees of XML or CIB data. + */ + return rc; + } /* Unpack constraints if any section will need them * (tickets may be referenced in constraints but not granted yet, From 20e97f00bbfcdc6ae61d5ab1150bd4eca3ac69a4 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Tue, 19 Nov 2024 10:36:25 -0500 Subject: [PATCH 04/11] Refactor: daemons: Don't continue if pcmk_unpack_scheduler_input fails. Also, there's no need to do any error reporting. pcmk__config_err will have already called crm_err in this case. --- daemons/execd/cts-exec-helper.c | 9 ++++++++- 1 file changed, 8 insertions(+), 1 deletion(-) diff --git a/daemons/execd/cts-exec-helper.c b/daemons/execd/cts-exec-helper.c index 1c4c1393e13..db056f59b16 100644 --- a/daemons/execd/cts-exec-helper.c +++ b/daemons/execd/cts-exec-helper.c @@ -475,7 +475,14 @@ generate_params(void) pcmk__set_scheduler_flags(scheduler, pcmk__sched_no_counts); scheduler->input = cib_xml_copy; scheduler->priv->now = crm_time_new(NULL); - pcmk_unpack_scheduler_input(scheduler); + + rc = pcmk_unpack_scheduler_input(scheduler); + + if (rc != pcmk_rc_ok) { + /* pcmk_unpack_scheduler_input -> pcmk__config_err will already log an error */ + pe_free_working_set(scheduler); + return rc; + } // Find resource in CIB rsc = pe_find_resource_with_flags(scheduler->priv->resources, From 9c405e1ca3820bb9cb5de3e4374a0295db8c947e Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 13:08:52 -0500 Subject: [PATCH 05/11] Low: tools: crm_simulate should error on too-new feature sets. The error message is hidden and only gets displayed if -V is given on the command line. Adding config error/warning handlers will cause the error to be displayed regardless. This could have been implemented in a couple ways, and there's tradeoffs here. I've chosen to duplicate what's happening in crm_verify, but instead of checking for verbosity (which is a global variable in that file), I'm checking out->is_quiet. This means that if you do `crm_simulate -Q`, you won't see the error message but you will get an error return code. This also means that `crm_simulate -Q -VVVV...`, you still won't see the error message. This may be a bug, but I'm not sure who would do that and I also think these sorts of problems are pervasive in our command line tools. Fix T521 --- tools/crm_simulate.c | 51 ++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 51 insertions(+) diff --git a/tools/crm_simulate.c b/tools/crm_simulate.c index df5359cb4f2..03969e709c2 100644 --- a/tools/crm_simulate.c +++ b/tools/crm_simulate.c @@ -342,6 +342,54 @@ static GOptionEntry source_entries[] = { { NULL } }; +/*! + * \internal + * \brief Output a configuration error + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_error(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "error: %s", buf); + } + va_end(ap); +} + +/*! + * \internal + * \brief Output a configuration warning + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_warning(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "warning: %s", buf); + } + va_end(ap); +} + static int setup_input(pcmk__output_t *out, const char *input, const char *output, GError **error) @@ -501,6 +549,9 @@ main(int argc, char **argv) out->quiet = args->quiet; + pcmk__set_config_error_handler(output_config_error, out); + pcmk__set_config_warning_handler(output_config_warning, out); + if (args->version) { out->version(out, false); goto done; From 6d1ecf60907a3c55c4bf9c2a93f04767f1471915 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 13:54:02 -0500 Subject: [PATCH 06/11] Low: tools: crm_resource should error on too-new feature sets. This is just like the previous patch for crm_simulate, complete with all the same problems regarding -Q and -V. --- tools/crm_resource.c | 70 ++++++++++++++++++++++++++++++++++++++++++-- 1 file changed, 68 insertions(+), 2 deletions(-) diff --git a/tools/crm_resource.c b/tools/crm_resource.c index bf9933cd09d..8c4a041a209 100644 --- a/tools/crm_resource.c +++ b/tools/crm_resource.c @@ -139,6 +139,54 @@ static pcmk__supported_format_t formats[] = { { NULL, NULL, NULL } }; +/*! + * \internal + * \brief Output a configuration error + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_error(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "error: %s", buf); + } + va_end(ap); +} + +/*! + * \internal + * \brief Output a configuration warning + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_warning(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "warning: %s", buf); + } + va_end(ap); +} + // Clean up and exit static crm_exit_t bye(crm_exit_t ec) @@ -959,7 +1007,15 @@ clear_constraints(pcmk__output_t *out) } scheduler->input = cib_xml; - pcmk_unpack_scheduler_input(scheduler); + rc = pcmk_unpack_scheduler_input(scheduler); + + if (rc != pcmk_rc_ok) { + /* Error printing is handled by pcmk__set_config_error_handler, and + * cleaning up scheduler is handled by the bye() function. + */ + g_list_free(before); + return rc; + } after = build_constraint_list(scheduler->input); remaining = pcmk__subtract_lists(before, after, (GCompareFunc) strcmp); @@ -993,7 +1049,14 @@ initialize_scheduler_data(xmlNode **cib_xml_orig) return rc; } - pcmk_unpack_scheduler_input(scheduler); + rc = pcmk_unpack_scheduler_input(scheduler); + if (rc != pcmk_rc_ok) { + /* Error printing is handled by pcmk__set_config_error_handler, and + * cleaning up scheduler is handled by the bye() function. + */ + return rc; + } + return pcmk_rc_ok; } @@ -1476,6 +1539,9 @@ main(int argc, char **argv) out->quiet = args->quiet; + pcmk__set_config_error_handler(output_config_error, out); + pcmk__set_config_warning_handler(output_config_warning, out); + crm_log_args(argc, argv); /* From dad03aded3e3dc843abf2bc838e15873d450ac22 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 14:07:54 -0500 Subject: [PATCH 07/11] Low: tools: crm_ticket should error on too-new feature sets. This is just like the previous patch to crm_simulate. However, one additional problem here is that it relies on using the deprecated -Q command line option. On the other hand, I think this is okay because we have a lot of work to do straightening out these sorts of options for all our command line tools. This is just one more thing we'll have to deal with at that time. --- tools/crm_ticket.c | 60 +++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 59 insertions(+), 1 deletion(-) diff --git a/tools/crm_ticket.c b/tools/crm_ticket.c index 7cb3701c1ef..d9319d56ee2 100644 --- a/tools/crm_ticket.c +++ b/tools/crm_ticket.c @@ -261,6 +261,54 @@ static GOptionEntry deprecated_entries[] = { { NULL } }; +/*! + * \internal + * \brief Output a configuration error + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_error(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "error: %s", buf); + } + va_end(ap); +} + +/*! + * \internal + * \brief Output a configuration warning + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_warning(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + pcmk__output_t *out = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!out->is_quiet(out)) { + out->err(out, "warning: %s", buf); + } + va_end(ap); +} + static void ticket_grant_warning(gchar *ticket_id) { @@ -382,6 +430,11 @@ main(int argc, char **argv) pe__register_messages(out); pcmk__register_lib_messages(out); + out->quiet = options.quiet; + + pcmk__set_config_error_handler(output_config_error, out); + pcmk__set_config_warning_handler(output_config_warning, out); + if (args->version) { out->version(out, false); goto done; @@ -441,7 +494,12 @@ main(int argc, char **argv) scheduler->input = cib_xml_copy; scheduler->priv->now = crm_time_new(NULL); - pcmk_unpack_scheduler_input(scheduler); + rc = pcmk_unpack_scheduler_input(scheduler); + if (rc != pcmk_rc_ok) { + /* Error printing is handled by pcmk__set_config_error_handler */ + exit_code = pcmk_rc2exitc(rc); + goto done; + } /* For recording the tickets that are referenced in PCMK_XE_RSC_TICKET * constraints but have never been granted yet. From 0e63dd2b61cf8224b06999caad252e0a33278ab3 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 15:24:38 -0500 Subject: [PATCH 08/11] Low: tools: crm_mon should error on too-new feature sets. This takes care of all callers of pcmk__output_cluster_status and pcmk__status. pcmk_status would also be affected, but at the moment there are no users of that function and anyway the config error handlers aren't public API. --- tools/crm_mon.c | 62 +++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 62 insertions(+) diff --git a/tools/crm_mon.c b/tools/crm_mon.c index 149ff7737e0..7441d358856 100644 --- a/tools/crm_mon.c +++ b/tools/crm_mon.c @@ -237,6 +237,59 @@ static void mon_st_callback_event(stonith_t * st, stonith_event_t * e); static void mon_st_callback_display(stonith_t * st, stonith_event_t * e); static void refresh_after_event(gboolean data_updated, gboolean enforce); +struct output_config_ctx { + pcmk__output_t *out; + bool quiet; +}; + +/*! + * \internal + * \brief Output a configuration error + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_error(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + struct output_config_ctx *occ = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!occ->quiet) { + occ->out->err(occ->out, "error: %s", buf); + } + va_end(ap); +} + +/*! + * \internal + * \brief Output a configuration warning + * + * \param[in] ctx Output object + * \param[in] msg printf(3)-style format string + * \param[in] ... Format string arguments + */ +G_GNUC_PRINTF(2, 3) +static void +output_config_warning(void *ctx, const char *msg, ...) +{ + va_list ap; + char *buf = NULL; + struct output_config_ctx *occ = ctx; + + va_start(ap, msg); + pcmk__assert(vasprintf(&buf, msg, ap) > 0); + if (!occ->quiet) { + occ->out->err(occ->out, "warning: %s", buf); + } + va_end(ap); +} + static uint32_t all_includes(mon_output_format_t fmt) { if ((fmt == mon_output_plain) || (fmt == mon_output_console)) { @@ -1408,6 +1461,7 @@ int main(int argc, char **argv) { int rc = pcmk_rc_ok; + struct output_config_ctx *ctx = NULL; GOptionGroup *output_group = NULL; args = pcmk__new_common_args(SUMMARY); @@ -1577,6 +1631,13 @@ main(int argc, char **argv) return clean_up(CRM_EX_OK); } + ctx = pcmk__assert_alloc(1, sizeof(struct output_config_ctx)); + ctx->out = out; + ctx->quiet = args->quiet; + + pcmk__set_config_error_handler(output_config_error, ctx); + pcmk__set_config_warning_handler(output_config_warning, ctx); + if (output_format == mon_output_xml) { show_opts |= pcmk_show_inactive_rscs | pcmk_show_timing; } @@ -1664,6 +1725,7 @@ main(int argc, char **argv) crm_info("Exiting %s", crm_system_name); + free(ctx); return clean_up(CRM_EX_OK); } From 52c9dce783f1d3c71e7508d7079800f0d94f4fe2 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 15:49:56 -0500 Subject: [PATCH 09/11] Refactor: libpacemaker: pcmk__schedule_actions should return a value. The point of this is to allow it to return the value from unpack_cib, which is returning the value from cluster_status. This allows us to check whether that function hit the too-new feature set CIB condition. --- lib/pacemaker/pcmk_scheduler.c | 26 +++++++++++++++++++------- 1 file changed, 19 insertions(+), 7 deletions(-) diff --git a/lib/pacemaker/pcmk_scheduler.c b/lib/pacemaker/pcmk_scheduler.c index b720749b31a..ba662b61309 100644 --- a/lib/pacemaker/pcmk_scheduler.c +++ b/lib/pacemaker/pcmk_scheduler.c @@ -734,14 +734,16 @@ log_unrunnable_actions(const pcmk_scheduler_t *scheduler) * \param[in,out] cib CIB XML to unpack (may be NULL if already unpacked) * \param[in] flags Scheduler flags to set in addition to defaults * \param[in,out] scheduler Scheduler data + * + * \return Standard Pacemaker return code */ -static void +static int unpack_cib(xmlNode *cib, unsigned long long flags, pcmk_scheduler_t *scheduler) { if (pcmk_is_set(scheduler->flags, pcmk__sched_have_status)) { crm_trace("Reusing previously calculated cluster status"); pcmk__set_scheduler_flags(scheduler, flags); - return; + return pcmk_rc_ok; } pcmk__assert(cib != NULL); @@ -756,7 +758,8 @@ unpack_cib(xmlNode *cib, unsigned long long flags, pcmk_scheduler_t *scheduler) pcmk__set_scheduler_flags(scheduler, flags); scheduler->input = cib; - pcmk_unpack_scheduler_input(scheduler); // Sets pcmk__sched_have_status + // Sets pcmk__sched_have_status + return pcmk_unpack_scheduler_input(scheduler); } /*! @@ -766,17 +769,24 @@ unpack_cib(xmlNode *cib, unsigned long long flags, pcmk_scheduler_t *scheduler) * \param[in,out] cib CIB XML to use as scheduler input * \param[in] flags Scheduler flags to set in addition to defaults * \param[in,out] scheduler Scheduler data + * + * \return Standard Pacemaker return code */ -void +int pcmk__schedule_actions(xmlNode *cib, unsigned long long flags, pcmk_scheduler_t *scheduler) { - unpack_cib(cib, flags, scheduler); + int rc = unpack_cib(cib, flags, scheduler); + + if (rc != pcmk_rc_ok) { + return rc; + } + pcmk__set_assignment_methods(scheduler); pcmk__apply_node_health(scheduler); pcmk__unpack_constraints(scheduler); if (pcmk_is_set(scheduler->flags, pcmk__sched_validate_only)) { - return; + return pcmk_rc_ok; } if (!pcmk_is_set(scheduler->flags, pcmk__sched_location_only) @@ -787,7 +797,7 @@ pcmk__schedule_actions(xmlNode *cib, unsigned long long flags, apply_node_criteria(scheduler); if (pcmk_is_set(scheduler->flags, pcmk__sched_location_only)) { - return; + return pcmk_rc_ok; } pcmk__create_internal_constraints(scheduler); @@ -808,6 +818,8 @@ pcmk__schedule_actions(xmlNode *cib, unsigned long long flags, if (get_crm_log_level() == LOG_TRACE) { log_unrunnable_actions(scheduler); } + + return pcmk_rc_ok; } /*! From c4247123f2eac718d05149768771c656c546620d Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Thu, 14 Nov 2024 16:33:04 -0500 Subject: [PATCH 10/11] Low: tools, libpacemaker: Check the return value of pcmk__schedule_actions. This takes care of most callers - the ones in the daemons are unlikely to be a problem. This allows catching the too-new schema condition in various other tools and displaying an error message to the user. Note that a couple other callers don't need to check the return value. I've added comments explaining why. --- include/pcmki/pcmki_scheduler.h | 4 ++-- lib/pacemaker/pcmk_simulate.c | 12 +++++++++++- lib/pacemaker/pcmk_verify.c | 5 ++++- tools/crm_resource_runtime.c | 17 +++++++++++++---- 4 files changed, 30 insertions(+), 8 deletions(-) diff --git a/include/pcmki/pcmki_scheduler.h b/include/pcmki/pcmki_scheduler.h index 77eb98ec8c1..cea22e57301 100644 --- a/include/pcmki/pcmki_scheduler.h +++ b/include/pcmki/pcmki_scheduler.h @@ -36,8 +36,8 @@ typedef struct { void pcmk__unpack_constraints(pcmk_scheduler_t *scheduler); -void pcmk__schedule_actions(xmlNode *cib, unsigned long long flags, - pcmk_scheduler_t *scheduler); +int pcmk__schedule_actions(xmlNode *cib, unsigned long long flags, + pcmk_scheduler_t *scheduler); GList *pcmk__copy_node_list(const GList *list, bool reset); diff --git a/lib/pacemaker/pcmk_simulate.c b/lib/pacemaker/pcmk_simulate.c index 35c4d215676..30d9c408e1d 100644 --- a/lib/pacemaker/pcmk_simulate.c +++ b/lib/pacemaker/pcmk_simulate.c @@ -362,6 +362,7 @@ profile_file(const char *xml_file, long long repeat, } for (int i = 0; i < repeat; ++i) { + int rc; xmlNode *input = cib_object; if (repeat > 1) { @@ -369,8 +370,12 @@ profile_file(const char *xml_file, long long repeat, } scheduler->input = input; set_effective_date(scheduler, false, use_date); - pcmk__schedule_actions(input, scheduler_flags, scheduler); + rc = pcmk__schedule_actions(input, scheduler_flags, scheduler); pe_reset_working_set(scheduler); + + if (rc != pcmk_rc_ok) { + break; + } } end = clock(); @@ -920,6 +925,11 @@ pcmk__simulate(pcmk_scheduler_t *scheduler, pcmk__output_t *out, scheduler->priv->out = logger_out; } + /* Likewise here - pcmk__schedule_actions only returns an error if + * cluster_status did, and there's nothing that could have changed since + * the first call to cause new errors here. So we don't need to check + * this return value either. + */ pcmk__schedule_actions(input, scheduler_flags, scheduler); if (logger_out == NULL) { diff --git a/lib/pacemaker/pcmk_verify.c b/lib/pacemaker/pcmk_verify.c index 893299c9ed4..36d307fe9b9 100644 --- a/lib/pacemaker/pcmk_verify.c +++ b/lib/pacemaker/pcmk_verify.c @@ -108,7 +108,10 @@ pcmk__verify(pcmk_scheduler_t *scheduler, pcmk__output_t *out, * ownership of the passed-in XML object, hence we pass in a copy * to the scheduler. */ - pcmk__schedule_actions(cib_object_copy, flags, scheduler); + rc = pcmk__schedule_actions(cib_object_copy, flags, scheduler); + if (rc != pcmk_rc_ok) { + pcmk__config_has_error = true; + } } verify_done: diff --git a/tools/crm_resource_runtime.c b/tools/crm_resource_runtime.c index b2c220fd872..fca17f851c0 100644 --- a/tools/crm_resource_runtime.c +++ b/tools/crm_resource_runtime.c @@ -1471,8 +1471,12 @@ update_dataset(cib_t *cib, pcmk_scheduler_t *scheduler, xmlNode **cib_xml_orig, goto done; } - pcmk__schedule_actions(scheduler->input, pcmk__sched_no_counts, - scheduler); + rc = pcmk__schedule_actions(scheduler->input, pcmk__sched_no_counts, + scheduler); + if (rc != pcmk_rc_ok) { + /* Error printing is handled by pcmk__set_config_error_handler */ + goto done; + } prev_quiet = out->is_quiet(out); out->quiet = true; @@ -2101,8 +2105,13 @@ wait_till_stable(pcmk__output_t *out, guint timeout_ms, cib_t * cib) if (rc != pcmk_rc_ok) { break; } - pcmk__schedule_actions(scheduler->input, pcmk__sched_no_counts, - scheduler); + + rc = pcmk__schedule_actions(scheduler->input, pcmk__sched_no_counts, + scheduler); + if (rc != pcmk_rc_ok) { + /* Error printing is handled by pcmk__set_config_error_handler */ + break; + } if (!printed_version_warning) { /* If the DC has a different version than the local node, the two From b4f4b751dc007a97d2c497ecd360c2e28dbf4306 Mon Sep 17 00:00:00 2001 From: Chris Lumens Date: Tue, 19 Nov 2024 11:01:53 -0500 Subject: [PATCH 11/11] Test: cts-cli: Update regression output for tool changes. * Remove the leading function name from various messages. This was most commonly "unpack_resources". * In the XML output format, move various messages from text that gets printed out to the XML output itself. This does end up with somewhat weird output with status="0" message="OK" followed by some error messages. * Add a couple warnings to crm_resource output. --- cts/cli/regression.acls.exp | 24 +- cts/cli/regression.crm_attribute.exp | 24 +- cts/cli/regression.crm_resource.exp | 217 ++++++++++-------- cts/cli/regression.validity.exp | 21 +- cts/scheduler/summary/797.summary | 1 + cts/scheduler/summary/bug-1822.summary | 4 + cts/scheduler/summary/bug-cl-5212.summary | 2 + cts/scheduler/summary/bug-lf-1852.summary | 10 + cts/scheduler/summary/bug-lf-2171.summary | 2 + cts/scheduler/summary/bug-lf-2606.summary | 8 + cts/scheduler/summary/bug-pm-11.summary | 1 + cts/scheduler/summary/bug-pm-12.summary | 1 + cts/scheduler/summary/bug-rh-1097457.summary | 4 + .../cancel-behind-moving-remote.summary | 192 ++++++++++++++++ .../summary/clone-anon-failcount.summary | 5 + .../summary/clone-anon-probe-1.summary | 6 + .../summary/clone-anon-probe-2.summary | 6 + .../summary/clone-require-all-1.summary | 1 + .../summary/clone-require-all-2.summary | 1 + .../summary/clone-require-all-3.summary | 1 + .../summary/clone-require-all-4.summary | 1 + .../summary/clone-require-all-5.summary | 1 + .../summary/clone-require-all-6.summary | 1 + .../summary/clone-require-all-7.summary | 1 + .../clone-require-all-no-interleave-1.summary | 1 + .../clone-require-all-no-interleave-2.summary | 1 + .../clone-require-all-no-interleave-3.summary | 1 + .../summary/coloc-clone-stays-active.summary | 1 + .../colocate-primitive-with-clone.summary | 3 + .../summary/colocation-influence.summary | 1 + .../summary/container-is-remote-node.summary | 3 + .../expire-non-blocked-failure.summary | 2 + cts/scheduler/summary/failcount-block.summary | 5 + .../summary/force-anon-clone-max.summary | 15 ++ .../summary/group-dependents.summary | 1 + .../summary/guest-host-not-fenceable.summary | 2 + cts/scheduler/summary/intervals.summary | 2 + .../summary/leftover-pending-monitor.summary | 1 + cts/scheduler/summary/novell-239079.summary | 9 + cts/scheduler/summary/novell-239082.summary | 12 + cts/scheduler/summary/novell-239087.summary | 10 + .../one-or-more-unrunnable-instances.summary | 1 + .../summary/order-serialize-set.summary | 2 + .../summary/order-wrong-kind.summary | 7 + .../summary/ordered-set-natural.summary | 1 + .../summary/priority-fencing-delay.summary | 6 + cts/scheduler/summary/promoted-9.summary | 2 + .../promoted-asymmetrical-order.summary | 16 ++ .../summary/promoted-failed-demote-2.summary | 3 + .../summary/promoted-failed-demote.summary | 3 + cts/scheduler/summary/promoted-group.summary | 5 + cts/scheduler/summary/promoted-notify.summary | 12 + .../summary/promoted-ordering.summary | 12 + .../summary/promoted-with-blocked.summary | 1 + cts/scheduler/summary/quorum-4.summary | 2 + cts/scheduler/summary/quorum-5.summary | 2 + cts/scheduler/summary/quorum-6.summary | 2 + cts/scheduler/summary/rec-node-10.summary | 1 + cts/scheduler/summary/rec-node-5.summary | 2 + cts/scheduler/summary/rec-node-8.summary | 1 + .../summary/remote-orphaned2.summary | 9 + .../summary/rsc-discovery-per-node.summary | 5 + .../summary/stop-failure-no-fencing.summary | 8 + .../summary/stop-failure-no-quorum.summary | 2 + .../summary/stop-failure-with-fencing.summary | 2 + cts/scheduler/summary/target-1.summary | 7 + cts/scheduler/summary/target-2.summary | 14 ++ .../summary/template-coloc-3.summary | 1 + .../summary/ticket-promoted-1.summary | 8 + .../summary/ticket-promoted-10.summary | 8 + .../summary/ticket-promoted-11.summary | 8 + .../summary/ticket-promoted-12.summary | 4 + .../summary/ticket-promoted-13.summary | 8 + .../summary/ticket-promoted-14.summary | 8 + .../summary/ticket-promoted-15.summary | 8 + .../summary/ticket-promoted-16.summary | 8 + .../summary/ticket-promoted-17.summary | 8 + .../summary/ticket-promoted-18.summary | 8 + .../summary/ticket-promoted-19.summary | 8 + .../summary/ticket-promoted-2.summary | 8 + .../summary/ticket-promoted-20.summary | 8 + .../summary/ticket-promoted-21.summary | 8 + .../summary/ticket-promoted-22.summary | 8 + .../summary/ticket-promoted-23.summary | 8 + .../summary/ticket-promoted-24.summary | 4 + .../summary/ticket-promoted-3.summary | 8 + .../summary/ticket-promoted-4.summary | 8 + .../summary/ticket-promoted-5.summary | 8 + .../summary/ticket-promoted-6.summary | 8 + .../summary/ticket-promoted-7.summary | 8 + .../summary/ticket-promoted-8.summary | 8 + .../summary/ticket-promoted-9.summary | 8 + .../summary/ticket-rsc-sets-1.summary | 8 + .../summary/ticket-rsc-sets-10.summary | 8 + .../summary/ticket-rsc-sets-11.summary | 8 + .../summary/ticket-rsc-sets-12.summary | 8 + .../summary/ticket-rsc-sets-13.summary | 8 + .../summary/ticket-rsc-sets-14.summary | 8 + .../summary/ticket-rsc-sets-2.summary | 8 + .../summary/ticket-rsc-sets-3.summary | 8 + .../summary/ticket-rsc-sets-4.summary | 8 + .../summary/ticket-rsc-sets-5.summary | 8 + .../summary/ticket-rsc-sets-6.summary | 8 + .../summary/ticket-rsc-sets-7.summary | 8 + .../summary/ticket-rsc-sets-8.summary | 8 + .../summary/ticket-rsc-sets-9.summary | 8 + cts/scheduler/summary/unrunnable-1.summary | 1 + cts/scheduler/summary/unrunnable-2.summary | 1 + .../whitebox-imply-stop-on-fence.summary | 6 + 109 files changed, 916 insertions(+), 122 deletions(-) diff --git a/cts/cli/regression.acls.exp b/cts/cli/regression.acls.exp index be91b93455a..17ef44847af 100644 --- a/cts/cli/regression.acls.exp +++ b/cts/cli/regression.acls.exp @@ -1040,9 +1040,9 @@ crm_resource: Error performing operation: Insufficient privileges =#=#=#= End test: l33t-haxor: Remove a resource meta attribute - Insufficient privileges (4) =#=#=#= * Passed: crm_resource - l33t-haxor: Remove a resource meta attribute =#=#=#= Begin test: niceguy: Create a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity pcmk__apply_creation_acl trace: Creation of scaffolding with id="dummy-meta_attributes" is implicitly allowed pcmk__apply_creation_acl trace: ACLs allow creation of with id="dummy-meta_attributes-target-role" Set 'dummy' option: id=dummy-meta_attributes-target-role set=dummy-meta_attributes name=target-role value=Stopped @@ -1128,9 +1128,9 @@ Set 'dummy' option: id=dummy-meta_attributes-target-role set=dummy-meta_attribut =#=#=#= End test: niceguy: Create a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - niceguy: Create a resource meta attribute =#=#=#= Begin test: niceguy: Query a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Stopped =#=#=#= Current cib after: niceguy: Query a resource meta attribute =#=#=#= @@ -1214,9 +1214,9 @@ Stopped =#=#=#= End test: niceguy: Query a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - niceguy: Query a resource meta attribute =#=#=#= Begin test: niceguy: Remove a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Deleted 'dummy' option: id=dummy-meta_attributes-target-role name=target-role =#=#=#= Current cib after: niceguy: Remove a resource meta attribute =#=#=#= @@ -1298,9 +1298,9 @@ Deleted 'dummy' option: id=dummy-meta_attributes-target-role name=target-role =#=#=#= End test: niceguy: Remove a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - niceguy: Remove a resource meta attribute =#=#=#= Begin test: niceguy: Create a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity pcmk__apply_creation_acl trace: ACLs allow creation of with id="dummy-meta_attributes-target-role" Set 'dummy' option: id=dummy-meta_attributes-target-role set=dummy-meta_attributes name=target-role value=Started =#=#=#= Current cib after: niceguy: Create a resource meta attribute =#=#=#= diff --git a/cts/cli/regression.crm_attribute.exp b/cts/cli/regression.crm_attribute.exp index b2005095ba8..5d581153046 100644 --- a/cts/cli/regression.crm_attribute.exp +++ b/cts/cli/regression.crm_attribute.exp @@ -1333,15 +1333,15 @@ Deleted crm_config option: id=(null) name=cluster-delay =#=#=#= End test: Delete cluster option with -i - OK (0) =#=#=#= * Passed: crm_attribute - Delete cluster option with -i =#=#=#= Begin test: Create node1 and bring it online =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Current cluster status: * Full List of Resources: * No resources @@ -1601,9 +1601,9 @@ scope=status name=fail-count-foo value=3 =#=#=#= End test: Query a fail count - OK (0) =#=#=#= * Passed: crm_failcount - Query a fail count =#=#=#= Begin test: Show node attributes with crm_simulate =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Current cluster status: * Node List: * Online: [ node1 ] diff --git a/cts/cli/regression.crm_resource.exp b/cts/cli/regression.crm_resource.exp index 9859fe316dd..63280a1896f 100644 --- a/cts/cli/regression.crm_resource.exp +++ b/cts/cli/regression.crm_resource.exp @@ -855,9 +855,9 @@ crm_resource: --class, --agent, and --provider can only be used with --validate =#=#=#= End test: crm_resource given resource config with invalid action - Incorrect usage (64) =#=#=#= * Passed: crm_resource - crm_resource given resource config with invalid action =#=#=#= Begin test: Create a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Set 'dummy' option: id=dummy-meta_attributes-is-managed set=dummy-meta_attributes name=is-managed value=false =#=#=#= Current cib after: Create a resource meta attribute =#=#=#= @@ -886,9 +886,9 @@ Set 'dummy' option: id=dummy-meta_attributes-is-managed set=dummy-meta_attribute =#=#=#= End test: Create a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - Create a resource meta attribute =#=#=#= Begin test: Query a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity false =#=#=#= Current cib after: Query a resource meta attribute =#=#=#= @@ -917,9 +917,9 @@ false =#=#=#= End test: Query a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - Query a resource meta attribute =#=#=#= Begin test: Remove a resource meta attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Deleted 'dummy' option: id=dummy-meta_attributes-is-managed name=is-managed =#=#=#= Current cib after: Remove a resource meta attribute =#=#=#= @@ -946,9 +946,6 @@ Deleted 'dummy' option: id=dummy-meta_attributes-is-managed name=is-managed =#=#=#= End test: Remove a resource meta attribute - OK (0) =#=#=#= * Passed: crm_resource - Remove a resource meta attribute =#=#=#= Begin test: Create another resource meta attribute (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity @@ -957,38 +954,50 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= End test: Create another resource meta attribute (XML) - OK (0) =#=#=#= * Passed: crm_resource - Create another resource meta attribute (XML) =#=#=#= Begin test: Show why a resource is not running (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= End test: Show why a resource is not running (XML) - OK (0) =#=#=#= * Passed: crm_resource - Show why a resource is not running (XML) =#=#=#= Begin test: Remove another resource meta attribute (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= End test: Remove another resource meta attribute (XML) - OK (0) =#=#=#= * Passed: crm_resource - Remove another resource meta attribute (XML) =#=#=#= Begin test: Get a non-existent attribute from a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity Attribute 'nonexistent' not found for 'dummy' @@ -996,9 +1005,9 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Get a non-existent attribute from a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Get a non-existent attribute from a resource element (XML) =#=#=#= Begin test: Get a non-existent attribute from a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Attribute 'nonexistent' not found for 'dummy' =#=#=#= Current cib after: Get a non-existent attribute from a resource element =#=#=#= @@ -1025,12 +1034,12 @@ Attribute 'nonexistent' not found for 'dummy' =#=#=#= End test: Get a non-existent attribute from a resource element - OK (0) =#=#=#= * Passed: crm_resource - Get a non-existent attribute from a resource element =#=#=#= Begin test: Get a non-existent attribute from a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity Attribute 'nonexistent' not found for 'dummy' @@ -1060,9 +1069,9 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Get a non-existent attribute from a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Get a non-existent attribute from a resource element (XML) =#=#=#= Begin test: Get an existent attribute from a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity ocf =#=#=#= Current cib after: Get an existent attribute from a resource element =#=#=#= @@ -1089,11 +1098,14 @@ ocf =#=#=#= End test: Get an existent attribute from a resource element - OK (0) =#=#=#= * Passed: crm_resource - Get an existent attribute from a resource element =#=#=#= Begin test: Set a non-existent attribute for a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= Current cib after: Set a non-existent attribute for a resource element (XML) =#=#=#= @@ -1120,11 +1132,14 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Set a non-existent attribute for a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Set a non-existent attribute for a resource element (XML) =#=#=#= Begin test: Set an existent attribute for a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= Current cib after: Set an existent attribute for a resource element (XML) =#=#=#= @@ -1151,11 +1166,14 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Set an existent attribute for a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Set an existent attribute for a resource element (XML) =#=#=#= Begin test: Delete an existent attribute for a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= Current cib after: Delete an existent attribute for a resource element (XML) =#=#=#= @@ -1182,11 +1200,14 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Delete an existent attribute for a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Delete an existent attribute for a resource element (XML) =#=#=#= Begin test: Delete a non-existent attribute for a resource element (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= Current cib after: Delete a non-existent attribute for a resource element (XML) =#=#=#= @@ -1213,9 +1234,9 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: Delete a non-existent attribute for a resource element (XML) - OK (0) =#=#=#= * Passed: crm_resource - Delete a non-existent attribute for a resource element (XML) =#=#=#= Begin test: Set a non-existent attribute for a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Set attribute: name=description value=test_description =#=#=#= Current cib after: Set a non-existent attribute for a resource element =#=#=#= @@ -1242,9 +1263,9 @@ Set attribute: name=description value=test_description =#=#=#= End test: Set a non-existent attribute for a resource element - OK (0) =#=#=#= * Passed: crm_resource - Set a non-existent attribute for a resource element =#=#=#= Begin test: Set an existent attribute for a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Set attribute: name=description value=test_description =#=#=#= Current cib after: Set an existent attribute for a resource element =#=#=#= @@ -1271,9 +1292,9 @@ Set attribute: name=description value=test_description =#=#=#= End test: Set an existent attribute for a resource element - OK (0) =#=#=#= * Passed: crm_resource - Set an existent attribute for a resource element =#=#=#= Begin test: Delete an existent attribute for a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Deleted attribute: description =#=#=#= Current cib after: Delete an existent attribute for a resource element =#=#=#= @@ -1300,9 +1321,9 @@ Deleted attribute: description =#=#=#= End test: Delete an existent attribute for a resource element - OK (0) =#=#=#= * Passed: crm_resource - Delete an existent attribute for a resource element =#=#=#= Begin test: Delete a non-existent attribute for a resource element =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Deleted attribute: description =#=#=#= Current cib after: Delete a non-existent attribute for a resource element =#=#=#= @@ -1329,9 +1350,9 @@ Deleted attribute: description =#=#=#= End test: Delete a non-existent attribute for a resource element - OK (0) =#=#=#= * Passed: crm_resource - Delete a non-existent attribute for a resource element =#=#=#= Begin test: Create a resource attribute =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Set 'dummy' option: id=dummy-instance_attributes-delay set=dummy-instance_attributes name=delay value=10s =#=#=#= Current cib after: Create a resource attribute =#=#=#= @@ -1361,9 +1382,9 @@ Set 'dummy' option: id=dummy-instance_attributes-delay set=dummy-instance_attrib =#=#=#= End test: Create a resource attribute - OK (0) =#=#=#= * Passed: crm_resource - Create a resource attribute =#=#=#= Begin test: List the configured resources =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Full List of Resources: * dummy (ocf:pacemaker:Dummy): Stopped =#=#=#= Current cib after: List the configured resources =#=#=#= @@ -1394,14 +1415,17 @@ Full List of Resources: =#=#=#= End test: List the configured resources - OK (0) =#=#=#= * Passed: crm_resource - List the configured resources =#=#=#= Begin test: List the configured resources (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= Current cib after: List the configured resources (XML) =#=#=#= @@ -1431,24 +1455,24 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure =#=#=#= End test: List the configured resources (XML) - OK (0) =#=#=#= * Passed: crm_resource - List the configured resources (XML) =#=#=#= Begin test: Implicitly list the configured resources =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Full List of Resources: * dummy (ocf:pacemaker:Dummy): Stopped =#=#=#= End test: Implicitly list the configured resources - OK (0) =#=#=#= * Passed: crm_resource - Implicitly list the configured resources =#=#=#= Begin test: List IDs of instantiated resources =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity dummy =#=#=#= End test: List IDs of instantiated resources - OK (0) =#=#=#= * Passed: crm_resource - List IDs of instantiated resources =#=#=#= Begin test: Show XML configuration of resource =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity dummy (ocf:pacemaker:Dummy): Stopped Resource XML: @@ -1460,9 +1484,6 @@ Resource XML: =#=#=#= End test: Show XML configuration of resource - OK (0) =#=#=#= * Passed: crm_resource - Show XML configuration of resource =#=#=#= Begin test: Show XML configuration of resource (XML) =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity @@ -1474,14 +1495,20 @@ unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure ]]> - + + + error: Resource start-up disabled since no STONITH resources have been defined + error: Either configure some or disable STONITH with the stonith-enabled option + error: NOTE: Clusters with shared data need STONITH to ensure data integrity + + =#=#=#= End test: Show XML configuration of resource (XML) - OK (0) =#=#=#= * Passed: crm_resource - Show XML configuration of resource (XML) =#=#=#= Begin test: Require a destination when migrating a resource that is stopped =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity crm_resource: Resource 'dummy' not moved: active in 0 locations. To prevent 'dummy' from running on a specific location, specify a node. =#=#=#= Current cib after: Require a destination when migrating a resource that is stopped =#=#=#= @@ -1512,9 +1539,9 @@ To prevent 'dummy' from running on a specific location, specify a node. =#=#=#= End test: Require a destination when migrating a resource that is stopped - Incorrect usage (64) =#=#=#= * Passed: crm_resource - Require a destination when migrating a resource that is stopped =#=#=#= Begin test: Don't support migration to non-existent locations =#=#=#= -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity crm_resource: Node 'i.do.not.exist' not found Error performing operation: No such object =#=#=#= Current cib after: Don't support migration to non-existent locations =#=#=#= @@ -1730,6 +1757,7 @@ WARNING: Creating rsc_location constraint 'cli-ban-dummy-on-node1' with a score =#=#=#= End test: Move a resource from its existing location - OK (0) =#=#=#= * Passed: crm_resource - Move a resource from its existing location =#=#=#= Begin test: Clear out constraints generated by --move =#=#=#= +warning: More than one node entry has name 'node1' Removing constraint: cli-ban-dummy-on-node1 =#=#=#= Current cib after: Clear out constraints generated by --move =#=#=#= @@ -2187,6 +2215,9 @@ Revised Cluster Status: =#=#=#= End test: Move dummy to node1 (XML) - OK (0) =#=#=#= * Passed: crm_resource - Move dummy to node1 (XML) =#=#=#= Begin test: Clear implicit constraints for dummy on node2 =#=#=#= +warning: More than one node entry has name 'node1' +warning: More than one node entry has name 'node2' +warning: More than one node entry has name 'node3' Removing constraint: cli-ban-dummy-on-node2 =#=#=#= Current cib after: Clear implicit constraints for dummy on node2 =#=#=#= @@ -3084,6 +3115,9 @@ WARNING: Creating rsc_location constraint 'cli-ban-dummy-on-node1' with a score =#=#=#= End test: Ban dummy from node1 for a short time - OK (0) =#=#=#= * Passed: crm_resource - Ban dummy from node1 for a short time =#=#=#= Begin test: Remove expired constraints =#=#=#= +warning: More than one node entry has name 'node1' +warning: More than one node entry has name 'node2' +warning: More than one node entry has name 'node3' Removing constraint: cli-ban-dummy-on-node1 =#=#=#= Current cib after: Remove expired constraints =#=#=#= @@ -3124,6 +3158,9 @@ Removing constraint: cli-ban-dummy-on-node1 =#=#=#= End test: Remove expired constraints - OK (0) =#=#=#= * Passed: sleep - Remove expired constraints =#=#=#= Begin test: Clear all implicit constraints for dummy =#=#=#= +warning: More than one node entry has name 'node1' +warning: More than one node entry has name 'node2' +warning: More than one node entry has name 'node3' Removing constraint: cli-prefer-dummy =#=#=#= Current cib after: Clear all implicit constraints for dummy =#=#=#= diff --git a/cts/cli/regression.validity.exp b/cts/cli/regression.validity.exp index c98b485ea26..3b70f24163f 100644 --- a/cts/cli/regression.validity.exp +++ b/cts/cli/regression.validity.exp @@ -66,15 +66,20 @@ Call failed: Update does not conform to the configured schema =#=#=#= End test: Set invalid rsc_order first-action value (schema validation disabled) - OK (0) =#=#=#= * Passed: cibadmin - Set invalid rsc_order first-action value (schema validation disabled) =#=#=#= Begin test: Run crm_simulate with invalid rsc_order first-action (schema validation disabled) =#=#=#= +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) Schema validation of configuration is disabled (support for validate-with set to "none" is deprecated and will be removed in a future release) -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity -invert_action warning: Unknown action 'break' specified in order constraint -invert_action warning: Unknown action 'break' specified in order constraint -unpack_resources error: Resource start-up disabled since no STONITH resources have been defined -unpack_resources error: Either configure some or disable STONITH with the stonith-enabled option -unpack_resources error: NOTE: Clusters with shared data need STONITH to ensure data integrity +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity +warning: Unknown action 'break' specified in order constraint +warning: Unknown action 'break' specified in order constraint +warning: Cannot invert constraint 'ord_1-2' (please specify inverse manually) +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Current cluster status: * Full List of Resources: * dummy1 (ocf:pacemaker:Dummy): Stopped diff --git a/cts/scheduler/summary/797.summary b/cts/scheduler/summary/797.summary index d31572ba3db..3618f487d6b 100644 --- a/cts/scheduler/summary/797.summary +++ b/cts/scheduler/summary/797.summary @@ -14,6 +14,7 @@ Current cluster status: * child_DoFencing:1 (stonith:ssh): Started c001n02 * child_DoFencing:2 (stonith:ssh): Started c001n03 * child_DoFencing:3 (stonith:ssh): Stopped +warning: Node c001n08 is unclean but cannot be fenced Transition Summary: * Stop DcIPaddr ( c001n03 ) due to no quorum diff --git a/cts/scheduler/summary/bug-1822.summary b/cts/scheduler/summary/bug-1822.summary index 3890a02730a..83b96772759 100644 --- a/cts/scheduler/summary/bug-1822.summary +++ b/cts/scheduler/summary/bug-1822.summary @@ -1,3 +1,4 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) Current cluster status: * Node List: * Online: [ process1a process2b ] @@ -10,6 +11,9 @@ Current cluster status: * Resource Group: ms-sf_group:1: * promotable_Stateful:1 (ocf:heartbeat:Dummy-statful): Promoted process1a * promotable_procdctl:1 (ocf:heartbeat:procdctl): Promoted process1a +error: Resetting 'on-fail' for promotable_Stateful:0 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for promotable_Stateful:1 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for promotable_procdctl:1 stop action to default value because 'stop' is not allowed for stop Transition Summary: * Stop promotable_Stateful:1 ( Promoted process1a ) due to node availability diff --git a/cts/scheduler/summary/bug-cl-5212.summary b/cts/scheduler/summary/bug-cl-5212.summary index 7cbe97558b6..496c0649892 100644 --- a/cts/scheduler/summary/bug-cl-5212.summary +++ b/cts/scheduler/summary/bug-cl-5212.summary @@ -19,6 +19,8 @@ Current cluster status: * prmPingd (ocf:pacemaker:ping): Started srv02 (UNCLEAN) * prmPingd (ocf:pacemaker:ping): Started srv01 (UNCLEAN) * Started: [ srv03 ] +warning: Node srv01 is unclean but cannot be fenced +warning: Node srv02 is unclean but cannot be fenced Transition Summary: * Stop prmStonith1-1 ( srv02 ) blocked diff --git a/cts/scheduler/summary/bug-lf-1852.summary b/cts/scheduler/summary/bug-lf-1852.summary index 26c73e166a5..bc8239c7637 100644 --- a/cts/scheduler/summary/bug-lf-1852.summary +++ b/cts/scheduler/summary/bug-lf-1852.summary @@ -1,3 +1,8 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ mysql-01 mysql-02 ] @@ -25,6 +30,11 @@ Executing Cluster Transition: * Resource action: drbd0:0 notify on mysql-02 * Resource action: drbd0:1 notify on mysql-01 * Pseudo action: ms-drbd0_confirmed-post_notify_running_0 +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/bug-lf-2171.summary b/cts/scheduler/summary/bug-lf-2171.summary index 5117608a20c..b1bd1b99c2d 100644 --- a/cts/scheduler/summary/bug-lf-2171.summary +++ b/cts/scheduler/summary/bug-lf-2171.summary @@ -1,3 +1,5 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) +warning: Support for the 'collocated' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) 2 of 4 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: diff --git a/cts/scheduler/summary/bug-lf-2606.summary b/cts/scheduler/summary/bug-lf-2606.summary index e0b7ebf0e68..9831385949b 100644 --- a/cts/scheduler/summary/bug-lf-2606.summary +++ b/cts/scheduler/summary/bug-lf-2606.summary @@ -12,6 +12,14 @@ Current cluster status: * Clone Set: ms3 [rsc3] (promotable): * Promoted: [ node2 ] * Unpromoted: [ node1 ] +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-unpromoted-5 is duplicate of rsc3-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Fence (reboot) node2 'rsc1 failed there' diff --git a/cts/scheduler/summary/bug-pm-11.summary b/cts/scheduler/summary/bug-pm-11.summary index c3f8f5b3af0..37f327fed94 100644 --- a/cts/scheduler/summary/bug-pm-11.summary +++ b/cts/scheduler/summary/bug-pm-11.summary @@ -1,3 +1,4 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) Current cluster status: * Node List: * Online: [ node-a node-b ] diff --git a/cts/scheduler/summary/bug-pm-12.summary b/cts/scheduler/summary/bug-pm-12.summary index 8defffe8d68..9f82560b3fb 100644 --- a/cts/scheduler/summary/bug-pm-12.summary +++ b/cts/scheduler/summary/bug-pm-12.summary @@ -1,3 +1,4 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) Current cluster status: * Node List: * Online: [ node-a node-b ] diff --git a/cts/scheduler/summary/bug-rh-1097457.summary b/cts/scheduler/summary/bug-rh-1097457.summary index f68a509609b..0b0f14e1222 100644 --- a/cts/scheduler/summary/bug-rh-1097457.summary +++ b/cts/scheduler/summary/bug-rh-1097457.summary @@ -32,6 +32,10 @@ Current cluster status: * FAKE4-IP (ocf:heartbeat:IPaddr2): Started lamaVM2 * Clone Set: FAKE6-clone [FAKE6]: * Started: [ lamaVM1 lamaVM2 lamaVM3 ] +warning: Invalid ordering constraint between FSlun4 and VM3 +warning: Invalid ordering constraint between FSlun3 and VM2 +warning: Invalid ordering constraint between FSlun2 and VM1 +warning: Invalid ordering constraint between FSlun1 and VM1 Transition Summary: * Fence (reboot) lamaVM2 (resource: VM2) 'guest is unclean' diff --git a/cts/scheduler/summary/cancel-behind-moving-remote.summary b/cts/scheduler/summary/cancel-behind-moving-remote.summary index 945f3c81da5..fd60a855d42 100644 --- a/cts/scheduler/summary/cancel-behind-moving-remote.summary +++ b/cts/scheduler/summary/cancel-behind-moving-remote.summary @@ -1,3 +1,99 @@ +warning: compute-0 requires fencing but fencing is disabled +warning: compute-1 requires fencing but fencing is disabled +warning: galera-bundle requires fencing but fencing is disabled +warning: galera-bundle-master requires fencing but fencing is disabled +warning: galera:0 requires fencing but fencing is disabled +warning: galera:1 requires fencing but fencing is disabled +warning: galera:2 requires fencing but fencing is disabled +warning: galera-bundle-podman-0 requires fencing but fencing is disabled +warning: galera-bundle-0 requires fencing but fencing is disabled +warning: galera-bundle-podman-1 requires fencing but fencing is disabled +warning: galera-bundle-1 requires fencing but fencing is disabled +warning: galera-bundle-podman-2 requires fencing but fencing is disabled +warning: galera-bundle-2 requires fencing but fencing is disabled +warning: rabbitmq-bundle requires fencing but fencing is disabled +warning: rabbitmq-bundle-clone requires fencing but fencing is disabled +warning: rabbitmq:0 requires fencing but fencing is disabled +warning: rabbitmq:1 requires fencing but fencing is disabled +warning: rabbitmq:2 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-0 requires fencing but fencing is disabled +warning: rabbitmq-bundle-0 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-1 requires fencing but fencing is disabled +warning: rabbitmq-bundle-1 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-2 requires fencing but fencing is disabled +warning: rabbitmq-bundle-2 requires fencing but fencing is disabled +warning: redis-bundle requires fencing but fencing is disabled +warning: redis-bundle-master requires fencing but fencing is disabled +warning: redis:0 requires fencing but fencing is disabled +warning: redis:1 requires fencing but fencing is disabled +warning: redis:2 requires fencing but fencing is disabled +warning: redis-bundle-podman-0 requires fencing but fencing is disabled +warning: redis-bundle-0 requires fencing but fencing is disabled +warning: redis-bundle-podman-1 requires fencing but fencing is disabled +warning: redis-bundle-1 requires fencing but fencing is disabled +warning: redis-bundle-podman-2 requires fencing but fencing is disabled +warning: redis-bundle-2 requires fencing but fencing is disabled +warning: ip-192.168.24.150 requires fencing but fencing is disabled +warning: ip-10.0.0.150 requires fencing but fencing is disabled +warning: ip-172.17.1.151 requires fencing but fencing is disabled +warning: ip-172.17.1.150 requires fencing but fencing is disabled +warning: ip-172.17.3.150 requires fencing but fencing is disabled +warning: ip-172.17.4.150 requires fencing but fencing is disabled +warning: haproxy-bundle requires fencing but fencing is disabled +warning: haproxy-bundle-podman-0 requires fencing but fencing is disabled +warning: haproxy-bundle-podman-1 requires fencing but fencing is disabled +warning: haproxy-bundle-podman-2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle requires fencing but fencing is disabled +warning: ovn-dbs-bundle-master requires fencing but fencing is disabled +warning: ovndb_servers:0 requires fencing but fencing is disabled +warning: ovndb_servers:1 requires fencing but fencing is disabled +warning: ovndb_servers:2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-0 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-0 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-1 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-1 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-2 requires fencing but fencing is disabled +warning: ip-172.17.1.87 requires fencing but fencing is disabled +warning: stonith-fence_compute-fence-nova requires fencing but fencing is disabled +warning: compute-unfence-trigger-clone requires fencing but fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:0 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:1 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:2 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:3 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:4 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:5 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:6 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:7 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:8 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:9 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:10 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:11 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:12 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:13 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:14 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:15 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:16 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:17 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:18 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:19 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:20 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:21 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:22 to "quorum" because fencing is disabled +warning: nova-evacuate requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400aa1373 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400dc23e0 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540040bb56 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400addd38 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540078fb07 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400ea59b0 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400066e50 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400e1534e requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540060dbba requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400e018b6 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400c87cdb requires fencing but fencing is disabled +warning: openstack-cinder-volume requires fencing but fencing is disabled +warning: openstack-cinder-volume-podman-0 requires fencing but fencing is disabled Using the original execution date of: 2021-02-15 01:40:51Z Current cluster status: * Node List: @@ -130,6 +226,102 @@ Executing Cluster Transition: * Pseudo action: ovn-dbs-bundle_promoted_0 * Resource action: ovndb_servers monitor=10000 on ovn-dbs-bundle-2 * Resource action: ovndb_servers monitor=30000 on ovn-dbs-bundle-0 +warning: compute-0 requires fencing but fencing is disabled +warning: compute-1 requires fencing but fencing is disabled +warning: galera-bundle requires fencing but fencing is disabled +warning: galera-bundle-master requires fencing but fencing is disabled +warning: galera:0 requires fencing but fencing is disabled +warning: galera:1 requires fencing but fencing is disabled +warning: galera:2 requires fencing but fencing is disabled +warning: galera-bundle-podman-0 requires fencing but fencing is disabled +warning: galera-bundle-0 requires fencing but fencing is disabled +warning: galera-bundle-podman-1 requires fencing but fencing is disabled +warning: galera-bundle-1 requires fencing but fencing is disabled +warning: galera-bundle-podman-2 requires fencing but fencing is disabled +warning: galera-bundle-2 requires fencing but fencing is disabled +warning: rabbitmq-bundle requires fencing but fencing is disabled +warning: rabbitmq-bundle-clone requires fencing but fencing is disabled +warning: rabbitmq:0 requires fencing but fencing is disabled +warning: rabbitmq:1 requires fencing but fencing is disabled +warning: rabbitmq:2 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-0 requires fencing but fencing is disabled +warning: rabbitmq-bundle-0 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-1 requires fencing but fencing is disabled +warning: rabbitmq-bundle-1 requires fencing but fencing is disabled +warning: rabbitmq-bundle-podman-2 requires fencing but fencing is disabled +warning: rabbitmq-bundle-2 requires fencing but fencing is disabled +warning: redis-bundle requires fencing but fencing is disabled +warning: redis-bundle-master requires fencing but fencing is disabled +warning: redis:0 requires fencing but fencing is disabled +warning: redis:1 requires fencing but fencing is disabled +warning: redis:2 requires fencing but fencing is disabled +warning: redis-bundle-podman-0 requires fencing but fencing is disabled +warning: redis-bundle-0 requires fencing but fencing is disabled +warning: redis-bundle-podman-1 requires fencing but fencing is disabled +warning: redis-bundle-1 requires fencing but fencing is disabled +warning: redis-bundle-podman-2 requires fencing but fencing is disabled +warning: redis-bundle-2 requires fencing but fencing is disabled +warning: ip-192.168.24.150 requires fencing but fencing is disabled +warning: ip-10.0.0.150 requires fencing but fencing is disabled +warning: ip-172.17.1.151 requires fencing but fencing is disabled +warning: ip-172.17.1.150 requires fencing but fencing is disabled +warning: ip-172.17.3.150 requires fencing but fencing is disabled +warning: ip-172.17.4.150 requires fencing but fencing is disabled +warning: haproxy-bundle requires fencing but fencing is disabled +warning: haproxy-bundle-podman-0 requires fencing but fencing is disabled +warning: haproxy-bundle-podman-1 requires fencing but fencing is disabled +warning: haproxy-bundle-podman-2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle requires fencing but fencing is disabled +warning: ovn-dbs-bundle-master requires fencing but fencing is disabled +warning: ovndb_servers:0 requires fencing but fencing is disabled +warning: ovndb_servers:1 requires fencing but fencing is disabled +warning: ovndb_servers:2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-0 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-0 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-1 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-1 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-podman-2 requires fencing but fencing is disabled +warning: ovn-dbs-bundle-2 requires fencing but fencing is disabled +warning: ip-172.17.1.87 requires fencing but fencing is disabled +warning: stonith-fence_compute-fence-nova requires fencing but fencing is disabled +warning: compute-unfence-trigger-clone requires fencing but fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:0 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:1 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:2 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:3 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:4 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:5 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:6 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:7 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:8 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:9 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:10 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:11 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:12 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:13 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:14 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:15 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:16 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:17 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:18 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:19 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:20 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:21 to "quorum" because fencing is disabled +warning: Resetting "requires" for compute-unfence-trigger:22 to "quorum" because fencing is disabled +warning: nova-evacuate requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400aa1373 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400dc23e0 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540040bb56 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400addd38 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540078fb07 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400ea59b0 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400066e50 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400e1534e requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-52540060dbba requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400e018b6 requires fencing but fencing is disabled +warning: stonith-fence_ipmilan-525400c87cdb requires fencing but fencing is disabled +warning: openstack-cinder-volume requires fencing but fencing is disabled +warning: openstack-cinder-volume-podman-0 requires fencing but fencing is disabled Using the original execution date of: 2021-02-15 01:40:51Z Revised Cluster Status: diff --git a/cts/scheduler/summary/clone-anon-failcount.summary b/cts/scheduler/summary/clone-anon-failcount.summary index 8d4f369e3e1..2b39b0b6874 100644 --- a/cts/scheduler/summary/clone-anon-failcount.summary +++ b/cts/scheduler/summary/clone-anon-failcount.summary @@ -36,6 +36,11 @@ Current cluster status: * Started: [ srv01 srv02 srv03 srv04 ] * Clone Set: clnG3dummy2 [clnG3dummy02]: * Started: [ srv01 srv02 srv03 srv04 ] +error: Resetting 'on-fail' for UmDummy01 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:0 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:1 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:2 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:3 stop action to default value because 'stop' is not allowed for stop Transition Summary: * Move UmVIPcheck ( srv01 -> srv04 ) diff --git a/cts/scheduler/summary/clone-anon-probe-1.summary b/cts/scheduler/summary/clone-anon-probe-1.summary index 51cf914a004..5539042553c 100644 --- a/cts/scheduler/summary/clone-anon-probe-1.summary +++ b/cts/scheduler/summary/clone-anon-probe-1.summary @@ -1,3 +1,6 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ mysql-01 mysql-02 ] @@ -17,6 +20,9 @@ Executing Cluster Transition: * Resource action: drbd0:0 start on mysql-01 * Resource action: drbd0:1 start on mysql-02 * Pseudo action: ms-drbd0_running_0 +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/clone-anon-probe-2.summary b/cts/scheduler/summary/clone-anon-probe-2.summary index 79a2fb8785e..aa37f7a828f 100644 --- a/cts/scheduler/summary/clone-anon-probe-2.summary +++ b/cts/scheduler/summary/clone-anon-probe-2.summary @@ -1,3 +1,6 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ mysql-01 mysql-02 ] @@ -14,6 +17,9 @@ Executing Cluster Transition: * Pseudo action: ms-drbd0_start_0 * Resource action: drbd0:1 start on mysql-01 * Pseudo action: ms-drbd0_running_0 +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/clone-require-all-1.summary b/cts/scheduler/summary/clone-require-all-1.summary index 7037eb8caa1..cf4274b2fb0 100644 --- a/cts/scheduler/summary/clone-require-all-1.summary +++ b/cts/scheduler/summary/clone-require-all-1.summary @@ -9,6 +9,7 @@ Current cluster status: * Stopped: [ rhel7-auto3 rhel7-auto4 ] * Clone Set: B-clone [B]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start B:0 ( rhel7-auto3 ) diff --git a/cts/scheduler/summary/clone-require-all-2.summary b/cts/scheduler/summary/clone-require-all-2.summary index 72d6f243f65..676810d22db 100644 --- a/cts/scheduler/summary/clone-require-all-2.summary +++ b/cts/scheduler/summary/clone-require-all-2.summary @@ -11,6 +11,7 @@ Current cluster status: * Stopped: [ rhel7-auto3 rhel7-auto4 ] * Clone Set: B-clone [B]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Move shooter ( rhel7-auto1 -> rhel7-auto3 ) diff --git a/cts/scheduler/summary/clone-require-all-3.summary b/cts/scheduler/summary/clone-require-all-3.summary index b828bffce28..485595407a3 100644 --- a/cts/scheduler/summary/clone-require-all-3.summary +++ b/cts/scheduler/summary/clone-require-all-3.summary @@ -12,6 +12,7 @@ Current cluster status: * Clone Set: B-clone [B]: * Started: [ rhel7-auto3 rhel7-auto4 ] * Stopped: [ rhel7-auto1 rhel7-auto2 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Move shooter ( rhel7-auto1 -> rhel7-auto3 ) diff --git a/cts/scheduler/summary/clone-require-all-4.summary b/cts/scheduler/summary/clone-require-all-4.summary index ebd7b6bb467..2632aebbec4 100644 --- a/cts/scheduler/summary/clone-require-all-4.summary +++ b/cts/scheduler/summary/clone-require-all-4.summary @@ -11,6 +11,7 @@ Current cluster status: * Clone Set: B-clone [B]: * Started: [ rhel7-auto3 rhel7-auto4 ] * Stopped: [ rhel7-auto1 rhel7-auto2 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Move shooter ( rhel7-auto1 -> rhel7-auto2 ) diff --git a/cts/scheduler/summary/clone-require-all-5.summary b/cts/scheduler/summary/clone-require-all-5.summary index b47049e8831..cae968b1ebb 100644 --- a/cts/scheduler/summary/clone-require-all-5.summary +++ b/cts/scheduler/summary/clone-require-all-5.summary @@ -9,6 +9,7 @@ Current cluster status: * Stopped: [ rhel7-auto3 rhel7-auto4 ] * Clone Set: B-clone [B]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start A:2 ( rhel7-auto3 ) diff --git a/cts/scheduler/summary/clone-require-all-6.summary b/cts/scheduler/summary/clone-require-all-6.summary index 5bae20c7285..ef1a99b2d30 100644 --- a/cts/scheduler/summary/clone-require-all-6.summary +++ b/cts/scheduler/summary/clone-require-all-6.summary @@ -10,6 +10,7 @@ Current cluster status: * Clone Set: B-clone [B]: * Started: [ rhel7-auto1 rhel7-auto3 rhel7-auto4 ] * Stopped: [ rhel7-auto2 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Stop A:0 ( rhel7-auto1 ) due to node availability diff --git a/cts/scheduler/summary/clone-require-all-7.summary b/cts/scheduler/summary/clone-require-all-7.summary index f0f2820c26d..ac4af30a846 100644 --- a/cts/scheduler/summary/clone-require-all-7.summary +++ b/cts/scheduler/summary/clone-require-all-7.summary @@ -8,6 +8,7 @@ Current cluster status: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] * Clone Set: B-clone [B]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start A:0 ( rhel7-auto2 ) diff --git a/cts/scheduler/summary/clone-require-all-no-interleave-1.summary b/cts/scheduler/summary/clone-require-all-no-interleave-1.summary index 646bfa3ef5c..50da4cc2166 100644 --- a/cts/scheduler/summary/clone-require-all-no-interleave-1.summary +++ b/cts/scheduler/summary/clone-require-all-no-interleave-1.summary @@ -11,6 +11,7 @@ Current cluster status: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] * Clone Set: C-clone [C]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start A:0 ( rhel7-auto3 ) diff --git a/cts/scheduler/summary/clone-require-all-no-interleave-2.summary b/cts/scheduler/summary/clone-require-all-no-interleave-2.summary index e40230cb527..bbd012cec27 100644 --- a/cts/scheduler/summary/clone-require-all-no-interleave-2.summary +++ b/cts/scheduler/summary/clone-require-all-no-interleave-2.summary @@ -11,6 +11,7 @@ Current cluster status: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] * Clone Set: C-clone [C]: * Stopped: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 rhel7-auto4 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start A:0 ( rhel7-auto4 ) diff --git a/cts/scheduler/summary/clone-require-all-no-interleave-3.summary b/cts/scheduler/summary/clone-require-all-no-interleave-3.summary index a22bf455b6a..85a03a0b378 100644 --- a/cts/scheduler/summary/clone-require-all-no-interleave-3.summary +++ b/cts/scheduler/summary/clone-require-all-no-interleave-3.summary @@ -14,6 +14,7 @@ Current cluster status: * Clone Set: C-clone [C]: * Started: [ rhel7-auto1 rhel7-auto2 rhel7-auto4 ] * Stopped: [ rhel7-auto3 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Move A:0 ( rhel7-auto4 -> rhel7-auto3 ) diff --git a/cts/scheduler/summary/coloc-clone-stays-active.summary b/cts/scheduler/summary/coloc-clone-stays-active.summary index cb212e1cde9..9e35a5d13a7 100644 --- a/cts/scheduler/summary/coloc-clone-stays-active.summary +++ b/cts/scheduler/summary/coloc-clone-stays-active.summary @@ -1,3 +1,4 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) 9 of 87 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: diff --git a/cts/scheduler/summary/colocate-primitive-with-clone.summary b/cts/scheduler/summary/colocate-primitive-with-clone.summary index e884428ee40..881ac31fb20 100644 --- a/cts/scheduler/summary/colocate-primitive-with-clone.summary +++ b/cts/scheduler/summary/colocate-primitive-with-clone.summary @@ -52,6 +52,9 @@ Current cluster status: * Clone Set: clnG3dummy2 [clnG3dummy02]: * Started: [ srv02 srv03 srv04 ] * Stopped: [ srv01 ] +error: Resetting 'on-fail' for clnG3dummy02:0 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:1 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for clnG3dummy02:2 stop action to default value because 'stop' is not allowed for stop Transition Summary: * Start UmVIPcheck ( srv04 ) diff --git a/cts/scheduler/summary/colocation-influence.summary b/cts/scheduler/summary/colocation-influence.summary index e240003d929..2cd66b670df 100644 --- a/cts/scheduler/summary/colocation-influence.summary +++ b/cts/scheduler/summary/colocation-influence.summary @@ -52,6 +52,7 @@ Current cluster status: * Promoted: [ rhel7-4 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-3 ] * Stopped: [ rhel7-5 ] +error: Constraint 'colocation-rsc1a-rsc1b-INFINITY' has invalid value for influence (using default) Transition Summary: * Move rsc1a ( rhel7-2 -> rhel7-3 ) diff --git a/cts/scheduler/summary/container-is-remote-node.summary b/cts/scheduler/summary/container-is-remote-node.summary index c022e896f44..a33c9ed7db1 100644 --- a/cts/scheduler/summary/container-is-remote-node.summary +++ b/cts/scheduler/summary/container-is-remote-node.summary @@ -24,6 +24,9 @@ Current cluster status: * FSdata1 (ocf:heartbeat:Filesystem): Started RNVM1 * RES1-IP (ocf:heartbeat:IPaddr2): Started RNVM1 * res-rsyslog (ocf:heartbeat:rsyslog.test): Started RNVM1 +warning: Invalid ordering constraint between gfs2-lv_1_1:0 and VM1 +warning: Invalid ordering constraint between clvmd:0 and VM1 +warning: Invalid ordering constraint between dlm:0 and VM1 Transition Summary: diff --git a/cts/scheduler/summary/expire-non-blocked-failure.summary b/cts/scheduler/summary/expire-non-blocked-failure.summary index 0ca6c540468..92ba7c8a82f 100644 --- a/cts/scheduler/summary/expire-non-blocked-failure.summary +++ b/cts/scheduler/summary/expire-non-blocked-failure.summary @@ -1,3 +1,4 @@ +warning: Ignoring failure timeout (1m) for rsc1 because it conflicts with on-fail=block 0 of 3 resource instances DISABLED and 1 BLOCKED from further action due to failure Current cluster status: @@ -13,6 +14,7 @@ Transition Summary: Executing Cluster Transition: * Cluster action: clear_failcount for rsc2 on node1 +warning: Ignoring failure timeout (1m) for rsc1 because it conflicts with on-fail=block Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/failcount-block.summary b/cts/scheduler/summary/failcount-block.summary index 646f76b400b..179497942da 100644 --- a/cts/scheduler/summary/failcount-block.summary +++ b/cts/scheduler/summary/failcount-block.summary @@ -1,3 +1,6 @@ +error: Ignoring invalid node_state entry without id +warning: Ignoring failure timeout (10s) for rsc_pcmk-2 because it conflicts with on-fail=block +warning: Ignoring failure timeout (10s) for rsc_pcmk-4 because it conflicts with on-fail=block 0 of 5 resource instances DISABLED and 1 BLOCKED from further action due to failure Current cluster status: @@ -25,6 +28,8 @@ Executing Cluster Transition: * Cluster action: clear_failcount for rsc_pcmk-5 on pcmk-1 * Resource action: rsc_pcmk-3 monitor=5000 on pcmk-1 * Resource action: rsc_pcmk-4 monitor=5000 on pcmk-1 +error: Ignoring invalid node_state entry without id +warning: Ignoring failure timeout (10s) for rsc_pcmk-2 because it conflicts with on-fail=block Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/force-anon-clone-max.summary b/cts/scheduler/summary/force-anon-clone-max.summary index d2320e9c571..2886410ab6e 100644 --- a/cts/scheduler/summary/force-anon-clone-max.summary +++ b/cts/scheduler/summary/force-anon-clone-max.summary @@ -1,3 +1,10 @@ +warning: Ignoring globally-unique for clone1 because lsb resources such as lsb1:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone1 because lsb resources such as lsb1:1 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:1 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:2 can be used only as anonymous clones +warning: Ignoring globally-unique for clone3 because lsb resources such as lsb3:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone3 because lsb resources such as lsb3:1 can be used only as anonymous clones Current cluster status: * Node List: * Online: [ node1 node2 node3 ] @@ -59,6 +66,14 @@ Executing Cluster Transition: * Resource action: dummy2:1 monitor=5000 on node2 * Resource action: lsb3:1 monitor=5000 on node2 * Pseudo action: clone3_running_0 +warning: Ignoring globally-unique for clone1 because lsb resources such as lsb1:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone1 because lsb resources such as lsb1:1 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:1 can be used only as anonymous clones +warning: Ignoring globally-unique for clone2 because lsb resources such as lsb2:2 can be used only as anonymous clones +warning: Ignoring globally-unique for clone3 because lsb resources such as lsb3:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone3 because lsb resources such as lsb3:1 can be used only as anonymous clones +warning: Ignoring globally-unique for clone3 because lsb resources such as lsb3:2 can be used only as anonymous clones Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/group-dependents.summary b/cts/scheduler/summary/group-dependents.summary index 33652555477..a8ce9c2915f 100644 --- a/cts/scheduler/summary/group-dependents.summary +++ b/cts/scheduler/summary/group-dependents.summary @@ -1,3 +1,4 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) Current cluster status: * Node List: * Online: [ asttest1 asttest2 ] diff --git a/cts/scheduler/summary/guest-host-not-fenceable.summary b/cts/scheduler/summary/guest-host-not-fenceable.summary index 9e3b5db405c..8fe32428bcd 100644 --- a/cts/scheduler/summary/guest-host-not-fenceable.summary +++ b/cts/scheduler/summary/guest-host-not-fenceable.summary @@ -18,6 +18,8 @@ Current cluster status: * stonith-fence_ipmilan-node1 (stonith:fence_ipmilan): Started node2 (UNCLEAN) * stonith-fence_ipmilan-node3 (stonith:fence_ipmilan): Started node2 (UNCLEAN) * stonith-fence_ipmilan-node2 (stonith:fence_ipmilan): Started node3 (UNCLEAN) +warning: Node node2 is unclean but cannot be fenced +warning: Node node3 is unclean but cannot be fenced Transition Summary: * Stop rabbitmq-bundle-docker-0 ( node1 ) due to no quorum diff --git a/cts/scheduler/summary/intervals.summary b/cts/scheduler/summary/intervals.summary index f6dc2e4b7fb..b4ebad3f69c 100644 --- a/cts/scheduler/summary/intervals.summary +++ b/cts/scheduler/summary/intervals.summary @@ -13,6 +13,8 @@ Current cluster status: * rsc4 (ocf:pacemaker:Dummy): FAILED rhel7-5 (blocked) * rsc5 (ocf:pacemaker:Dummy): Started rhel7-1 * rsc6 (ocf:pacemaker:Dummy): Started rhel7-2 +error: Operation rsc3-monitor-interval-P40S is duplicate of rsc3-monitor-interval-40s (do not use same name and interval combination more than once per resource) +error: Operation rsc3-monitor-interval-P40S is duplicate of rsc3-monitor-interval-40s (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc2 ( rhel7-3 ) diff --git a/cts/scheduler/summary/leftover-pending-monitor.summary b/cts/scheduler/summary/leftover-pending-monitor.summary index 04b03f29d85..d5e7e39f107 100644 --- a/cts/scheduler/summary/leftover-pending-monitor.summary +++ b/cts/scheduler/summary/leftover-pending-monitor.summary @@ -9,6 +9,7 @@ Current cluster status: * Clone Set: promotable-1 [stateful-1] (promotable): * Promoted: [ node-3 ] * Stopped: [ node-1 node-2 ] +warning: Support for the Promoted role is deprecated and will be removed in a future release. Use Promoted instead. Transition Summary: * Start stateful-1:1 ( node-1 ) due to unrunnable stateful-1:0 monitor (blocked) diff --git a/cts/scheduler/summary/novell-239079.summary b/cts/scheduler/summary/novell-239079.summary index 0afbba57970..401ccd11d76 100644 --- a/cts/scheduler/summary/novell-239079.summary +++ b/cts/scheduler/summary/novell-239079.summary @@ -1,3 +1,7 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ xen-1 xen-2 ] @@ -22,6 +26,11 @@ Executing Cluster Transition: * Resource action: drbd0:0 notify on xen-1 * Resource action: drbd0:1 notify on xen-2 * Pseudo action: ms-drbd0_confirmed-post_notify_running_0 +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/novell-239082.summary b/cts/scheduler/summary/novell-239082.summary index 051c0220e01..5d27e93076b 100644 --- a/cts/scheduler/summary/novell-239082.summary +++ b/cts/scheduler/summary/novell-239082.summary @@ -1,3 +1,8 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ xen-1 xen-2 ] @@ -7,6 +12,8 @@ Current cluster status: * Clone Set: ms-drbd0 [drbd0] (promotable): * Promoted: [ xen-1 ] * Unpromoted: [ xen-2 ] +warning: Support for setting meta-attributes (such as target_role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target_role) to the explicit value '#default' is deprecated and will be removed in a future release Transition Summary: * Move fs_1 ( xen-1 -> xen-2 ) @@ -47,6 +54,11 @@ Executing Cluster Transition: * Resource action: drbd0:0 notify on xen-2 * Pseudo action: ms-drbd0_confirmed-post_notify_promoted_0 * Resource action: fs_1 start on xen-2 +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/novell-239087.summary b/cts/scheduler/summary/novell-239087.summary index 0c158d3873d..df2db7abfb9 100644 --- a/cts/scheduler/summary/novell-239087.summary +++ b/cts/scheduler/summary/novell-239087.summary @@ -1,3 +1,7 @@ +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Current cluster status: * Node List: * Online: [ xen-1 xen-2 ] @@ -7,6 +11,12 @@ Current cluster status: * Clone Set: ms-drbd0 [drbd0] (promotable): * Promoted: [ xen-1 ] * Unpromoted: [ xen-2 ] +warning: Support for setting meta-attributes (such as target_role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target_role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release +warning: Support for setting meta-attributes (such as target-role) to the explicit value '#default' is deprecated and will be removed in a future release Transition Summary: diff --git a/cts/scheduler/summary/one-or-more-unrunnable-instances.summary b/cts/scheduler/summary/one-or-more-unrunnable-instances.summary index 58c572d199e..13eeacbffeb 100644 --- a/cts/scheduler/summary/one-or-more-unrunnable-instances.summary +++ b/cts/scheduler/summary/one-or-more-unrunnable-instances.summary @@ -132,6 +132,7 @@ Current cluster status: * mrg-07 (ocf:pacemaker:remote): Started rdo7-node1 * mrg-08 (ocf:pacemaker:remote): Started rdo7-node2 * mrg-09 (ocf:pacemaker:remote): Started rdo7-node3 +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start keystone:0 ( rdo7-node2 ) diff --git a/cts/scheduler/summary/order-serialize-set.summary b/cts/scheduler/summary/order-serialize-set.summary index b0b759b51ce..54fd7b18d44 100644 --- a/cts/scheduler/summary/order-serialize-set.summary +++ b/cts/scheduler/summary/order-serialize-set.summary @@ -14,6 +14,8 @@ Current cluster status: * edge (ocf:heartbeat:Xen): Started xen-a * base (ocf:heartbeat:Xen): Started xen-a * Email_Alerting (ocf:heartbeat:MailTo): Started xen-b +warning: Ignoring symmetrical for 'serialize-xen' because not valid with kind of 'Serialize' +warning: Ignoring symmetrical for 'xen-set' because not valid with kind of 'Serialize' Transition Summary: * Restart xen-a-fencing ( xen-b ) due to resource definition change diff --git a/cts/scheduler/summary/order-wrong-kind.summary b/cts/scheduler/summary/order-wrong-kind.summary index 903a25c7239..48c34546218 100644 --- a/cts/scheduler/summary/order-wrong-kind.summary +++ b/cts/scheduler/summary/order-wrong-kind.summary @@ -1,4 +1,7 @@ +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) Schema validation of configuration is disabled (support for validate-with set to "none" is deprecated and will be removed in a future release) +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) +warning: Support for validate-with='none' is deprecated and will be removed in a future release without the possibility of upgrades (manually edit to use a supported schema) Current cluster status: * Node List: * Online: [ node1 ] @@ -8,6 +11,10 @@ Current cluster status: * rsc2 (ocf:heartbeat:apache): Started node1 * rsc3 (ocf:heartbeat:apache): Stopped * rsc4 (ocf:heartbeat:apache): Started node1 +error: Resetting 'kind' for constraint order1 to 'Mandatory' because 'foo' is not valid +error: Resetting 'kind' for constraint order1 to 'Mandatory' because 'foo' is not valid +error: Resetting 'kind' for constraint order1 to 'Mandatory' because 'foo' is not valid +error: Resetting 'kind' for constraint order1 to 'Mandatory' because 'foo' is not valid Transition Summary: * Start rsc1 ( node1 ) diff --git a/cts/scheduler/summary/ordered-set-natural.summary b/cts/scheduler/summary/ordered-set-natural.summary index b944e0d6f40..bf96e250f78 100644 --- a/cts/scheduler/summary/ordered-set-natural.summary +++ b/cts/scheduler/summary/ordered-set-natural.summary @@ -21,6 +21,7 @@ Current cluster status: * dummy3-5 (ocf:heartbeat:Dummy): Stopped * dummy2-4 (ocf:heartbeat:Dummy): Stopped * dummy2-5 (ocf:heartbeat:Dummy): Stopped +warning: Support for 'ordering' other than 'group' in resource_set (such as pcs_rsc_set_dummy3-1_dummy3-2_dummy3-3_dummy3-4_dummy3-5-1) is deprecated and will be removed in a future release Transition Summary: * Start dummy1-1 ( node1 ) due to no quorum (blocked) diff --git a/cts/scheduler/summary/priority-fencing-delay.summary b/cts/scheduler/summary/priority-fencing-delay.summary index ce5aff2562f..0c6bc702f20 100644 --- a/cts/scheduler/summary/priority-fencing-delay.summary +++ b/cts/scheduler/summary/priority-fencing-delay.summary @@ -24,6 +24,12 @@ Current cluster status: * R-lxc-01_kiff-01 (ocf:heartbeat:VirtualDomain): FAILED kiff-01 (UNCLEAN) * R-lxc-02_kiff-01 (ocf:heartbeat:VirtualDomain): Started kiff-01 (UNCLEAN) * R-lxc-02_kiff-02 (ocf:heartbeat:VirtualDomain): Started kiff-02 +warning: Invalid ordering constraint between shared0:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between clvmd:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between dlm:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between shared0:0 and R-lxc-01_kiff-02 +warning: Invalid ordering constraint between clvmd:0 and R-lxc-01_kiff-02 +warning: Invalid ordering constraint between dlm:0 and R-lxc-01_kiff-02 Transition Summary: * Fence (reboot) lxc-02_kiff-01 (resource: R-lxc-02_kiff-01) 'guest is unclean' diff --git a/cts/scheduler/summary/promoted-9.summary b/cts/scheduler/summary/promoted-9.summary index 69dab46a2ce..7be9cf7c72c 100644 --- a/cts/scheduler/summary/promoted-9.summary +++ b/cts/scheduler/summary/promoted-9.summary @@ -29,6 +29,8 @@ Current cluster status: * ocf_msdummy:5 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped * ocf_msdummy:6 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped * ocf_msdummy:7 (ocf:heartbeat:/usr/lib64/heartbeat/cts/OCFMSDummy): Stopped +warning: Node sgi2 is unclean but cannot be fenced +warning: Node test02 is unclean but cannot be fenced Transition Summary: * Start DcIPaddr ( va1 ) due to no quorum (blocked) diff --git a/cts/scheduler/summary/promoted-asymmetrical-order.summary b/cts/scheduler/summary/promoted-asymmetrical-order.summary index 591ff18a04f..1702272f720 100644 --- a/cts/scheduler/summary/promoted-asymmetrical-order.summary +++ b/cts/scheduler/summary/promoted-asymmetrical-order.summary @@ -11,6 +11,22 @@ Current cluster status: * Clone Set: ms2 [rsc2] (promotable): * Promoted: [ node2 ] * Unpromoted: [ node1 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc2-monitor-unpromoted-5 is duplicate of rsc2-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1:0 ( Promoted node1 ) due to node availability diff --git a/cts/scheduler/summary/promoted-failed-demote-2.summary b/cts/scheduler/summary/promoted-failed-demote-2.summary index 3f317fabeaa..02f3ee7e670 100644 --- a/cts/scheduler/summary/promoted-failed-demote-2.summary +++ b/cts/scheduler/summary/promoted-failed-demote-2.summary @@ -10,6 +10,9 @@ Current cluster status: * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a +error: Resetting 'on-fail' for stateful-1:0 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for stateful-1:1 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for stateful-2:1 stop action to default value because 'stop' is not allowed for stop Transition Summary: * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability diff --git a/cts/scheduler/summary/promoted-failed-demote.summary b/cts/scheduler/summary/promoted-failed-demote.summary index 70b3e1b2cff..e9f1a1baa92 100644 --- a/cts/scheduler/summary/promoted-failed-demote.summary +++ b/cts/scheduler/summary/promoted-failed-demote.summary @@ -10,6 +10,9 @@ Current cluster status: * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a +error: Resetting 'on-fail' for stateful-1:0 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for stateful-1:1 stop action to default value because 'stop' is not allowed for stop +error: Resetting 'on-fail' for stateful-2:1 stop action to default value because 'stop' is not allowed for stop Transition Summary: * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability diff --git a/cts/scheduler/summary/promoted-group.summary b/cts/scheduler/summary/promoted-group.summary index 44b380c25b7..03a7f79afaf 100644 --- a/cts/scheduler/summary/promoted-group.summary +++ b/cts/scheduler/summary/promoted-group.summary @@ -1,3 +1,5 @@ +warning: Support for the 'ordered' group meta-attribute is deprecated and will be removed in a future release (use a resource set instead) +error: Resetting 'on-fail' for monitor of resource_1 to 'stop' because 'fence' is not valid when fencing is disabled Current cluster status: * Node List: * Online: [ rh44-1 rh44-2 ] @@ -10,6 +12,8 @@ Current cluster status: * promotable_Stateful:0 (ocf:heartbeat:Stateful): Unpromoted rh44-2 * Resource Group: grp_ms_sf:1: * promotable_Stateful:1 (ocf:heartbeat:Stateful): Unpromoted rh44-1 +error: Resetting 'on-fail' for stop of resource_1 to 'stop' because 'fence' is not valid when fencing is disabled +error: Resetting 'on-fail' for start of resource_1 to 'stop' because 'fence' is not valid when fencing is disabled Transition Summary: * Promote promotable_Stateful:1 ( Unpromoted -> Promoted rh44-1 ) @@ -22,6 +26,7 @@ Executing Cluster Transition: * Pseudo action: grp_ms_sf:1_promoted_0 * Resource action: promotable_Stateful:1 monitor=6000 on rh44-1 * Pseudo action: ms-sf_promoted_0 +error: Resetting 'on-fail' for monitor of resource_1 to 'stop' because 'fence' is not valid when fencing is disabled Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/promoted-notify.summary b/cts/scheduler/summary/promoted-notify.summary index f0fb04027d8..098e945dce2 100644 --- a/cts/scheduler/summary/promoted-notify.summary +++ b/cts/scheduler/summary/promoted-notify.summary @@ -6,6 +6,18 @@ Current cluster status: * shooter (stonith:fence_xvm): Started rhel7-auto1 * Clone Set: fake-master [fake] (promotable): * Unpromoted: [ rhel7-auto1 rhel7-auto2 rhel7-auto3 ] +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) +error: Operation fake-monitor-interval-10-role-Unpromoted is duplicate of fake-monitor-interval-10-role-Promoted (do not use same name and interval combination more than once per resource) Transition Summary: * Promote fake:0 ( Unpromoted -> Promoted rhel7-auto1 ) diff --git a/cts/scheduler/summary/promoted-ordering.summary b/cts/scheduler/summary/promoted-ordering.summary index 0ef1bd89e86..84158af223a 100644 --- a/cts/scheduler/summary/promoted-ordering.summary +++ b/cts/scheduler/summary/promoted-ordering.summary @@ -1,3 +1,5 @@ +warning: Ignoring globally-unique for clone_webservice because lsb resources such as mysql-proxy:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone_webservice because lsb resources such as mysql-proxy:1 can be used only as anonymous clones Current cluster status: * Node List: * Online: [ webcluster01 ] @@ -21,6 +23,14 @@ Current cluster status: * Clone Set: ms_drbd_mysql [drbd_mysql] (promotable): * Stopped: [ webcluster01 webcluster02 ] * fs_mysql (ocf:heartbeat:Filesystem): Stopped +warning: No resource, template, or tag named 'drbd_mysql' +error: Ignoring constraint 'colo_drbd_mysql_ip0' because 'drbd_mysql' is not a valid resource or tag +warning: No resource, template, or tag named 'drbd_mysql' +error: Ignoring constraint 'colo_drbd_mysql_ip1' because 'drbd_mysql' is not a valid resource or tag +warning: No resource, template, or tag named 'drbd_www' +error: Ignoring constraint 'colo_drbd_www_ip0' because 'drbd_www' is not a valid resource or tag +warning: No resource, template, or tag named 'drbd_www' +error: Ignoring constraint 'colo_drbd_www_ip1' because 'drbd_www' is not a valid resource or tag Transition Summary: * Start extip_1 ( webcluster01 ) @@ -68,6 +78,8 @@ Executing Cluster Transition: * Pseudo action: ms_drbd_www_confirmed-post_notify_running_0 * Resource action: drbd_mysql:0 notify on webcluster01 * Pseudo action: ms_drbd_mysql_confirmed-post_notify_running_0 +warning: Ignoring globally-unique for clone_webservice because lsb resources such as mysql-proxy:0 can be used only as anonymous clones +warning: Ignoring globally-unique for clone_webservice because lsb resources such as mysql-proxy:1 can be used only as anonymous clones Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/promoted-with-blocked.summary b/cts/scheduler/summary/promoted-with-blocked.summary index 82177a9a6a2..c38b1ce49fd 100644 --- a/cts/scheduler/summary/promoted-with-blocked.summary +++ b/cts/scheduler/summary/promoted-with-blocked.summary @@ -10,6 +10,7 @@ Current cluster status: * Clone Set: rsc2-clone [rsc2] (promotable): * Stopped: [ node1 node2 node3 node4 node5 ] * rsc3 (ocf:pacemaker:Dummy): Stopped (disabled) +warning: Support for the Promoted role is deprecated and will be removed in a future release. Use Promoted instead. Transition Summary: * Start rsc1 ( node2 ) due to unrunnable rsc3 start (blocked) diff --git a/cts/scheduler/summary/quorum-4.summary b/cts/scheduler/summary/quorum-4.summary index 3d0c88e81f6..0132adc92bb 100644 --- a/cts/scheduler/summary/quorum-4.summary +++ b/cts/scheduler/summary/quorum-4.summary @@ -6,6 +6,8 @@ Current cluster status: * Full List of Resources: * child_DoFencing (stonith:ssh): Stopped +warning: Node hadev1 is unclean but cannot be fenced +warning: Node hadev3 is unclean but cannot be fenced Transition Summary: * Start child_DoFencing ( hadev2 ) diff --git a/cts/scheduler/summary/quorum-5.summary b/cts/scheduler/summary/quorum-5.summary index 1e7abf38ee9..407dad631d6 100644 --- a/cts/scheduler/summary/quorum-5.summary +++ b/cts/scheduler/summary/quorum-5.summary @@ -8,6 +8,8 @@ Current cluster status: * Resource Group: group1: * child_DoFencing_1 (stonith:ssh): Stopped * child_DoFencing_2 (stonith:ssh): Stopped +warning: Node hadev1 is unclean but cannot be fenced +warning: Node hadev3 is unclean but cannot be fenced Transition Summary: * Start child_DoFencing_1 ( hadev2 ) diff --git a/cts/scheduler/summary/quorum-6.summary b/cts/scheduler/summary/quorum-6.summary index 321410d5b5e..04f41803b4e 100644 --- a/cts/scheduler/summary/quorum-6.summary +++ b/cts/scheduler/summary/quorum-6.summary @@ -14,6 +14,8 @@ Current cluster status: * child_DoFencing:5 (stonith:ssh): Stopped * child_DoFencing:6 (stonith:ssh): Stopped * child_DoFencing:7 (stonith:ssh): Stopped +warning: Node hadev1 is unclean but cannot be fenced +warning: Node hadev3 is unclean but cannot be fenced Transition Summary: * Start child_DoFencing:0 ( hadev2 ) diff --git a/cts/scheduler/summary/rec-node-10.summary b/cts/scheduler/summary/rec-node-10.summary index a77b2a14eed..2df3f57eb8d 100644 --- a/cts/scheduler/summary/rec-node-10.summary +++ b/cts/scheduler/summary/rec-node-10.summary @@ -7,6 +7,7 @@ Current cluster status: * stonith-1 (stonith:dummy): Stopped * rsc1 (ocf:heartbeat:apache): Started node1 (UNCLEAN) * rsc2 (ocf:heartbeat:apache): Started node1 (UNCLEAN) +warning: Node node1 is unclean but cannot be fenced Transition Summary: * Start stonith-1 ( node2 ) due to no quorum (blocked) diff --git a/cts/scheduler/summary/rec-node-5.summary b/cts/scheduler/summary/rec-node-5.summary index a4128ca167d..9ed88580a6c 100644 --- a/cts/scheduler/summary/rec-node-5.summary +++ b/cts/scheduler/summary/rec-node-5.summary @@ -6,6 +6,8 @@ Current cluster status: * Full List of Resources: * rsc1 (ocf:heartbeat:apache): Stopped * rsc2 (ocf:heartbeat:apache): Stopped +warning: Node node1 is unclean but cannot be fenced +warning: Resource functionality and data integrity cannot be guaranteed (configure, enable, and test fencing to correct this) Transition Summary: * Start rsc1 ( node2 ) diff --git a/cts/scheduler/summary/rec-node-8.summary b/cts/scheduler/summary/rec-node-8.summary index 226e333dfc1..c20908be574 100644 --- a/cts/scheduler/summary/rec-node-8.summary +++ b/cts/scheduler/summary/rec-node-8.summary @@ -8,6 +8,7 @@ Current cluster status: * rsc1 (ocf:heartbeat:apache): Started node1 (UNCLEAN) * rsc2 (ocf:heartbeat:apache): Started node1 (UNCLEAN) * rsc3 (ocf:heartbeat:apache): Stopped +warning: Node node1 is unclean but cannot be fenced Transition Summary: * Start stonith-1 ( node2 ) due to quorum freeze (blocked) diff --git a/cts/scheduler/summary/remote-orphaned2.summary b/cts/scheduler/summary/remote-orphaned2.summary index 9b0091467b0..f9e0c032425 100644 --- a/cts/scheduler/summary/remote-orphaned2.summary +++ b/cts/scheduler/summary/remote-orphaned2.summary @@ -1,3 +1,6 @@ +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Current cluster status: * Node List: * RemoteNode mrg-02: UNCLEAN (offline) @@ -10,6 +13,12 @@ Current cluster status: * libvirtd-compute (systemd:libvirtd): ORPHANED Started [ mrg-03 mrg-02 mrg-04 ] * ceilometer-compute (systemd:openstack-ceilometer-compute): ORPHANED Started [ mrg-03 mrg-02 mrg-04 ] * nova-compute (systemd:openstack-nova-compute): ORPHANED Started [ mrg-03 mrg-02 mrg-04 ] +warning: Node mrg-02 is unclean but cannot be fenced +warning: Node mrg-03 is unclean but cannot be fenced +warning: Node mrg-04 is unclean but cannot be fenced +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Transition Summary: diff --git a/cts/scheduler/summary/rsc-discovery-per-node.summary b/cts/scheduler/summary/rsc-discovery-per-node.summary index 3c34ced4ff1..150799f577b 100644 --- a/cts/scheduler/summary/rsc-discovery-per-node.summary +++ b/cts/scheduler/summary/rsc-discovery-per-node.summary @@ -1,3 +1,6 @@ +warning: Ignoring resource-discovery-enabled attribute for 18node1 because disabling resource discovery is not allowed for cluster nodes +warning: Ignoring resource-discovery-enabled attribute for 18node2 because disabling resource discovery is not allowed for cluster nodes +warning: Support for the resource-discovery-enabled node attribute is deprecated and will be removed (and behave as 'true') in a future release. Current cluster status: * Node List: * Online: [ 18builder 18node1 18node2 18node3 18node4 ] @@ -110,6 +113,8 @@ Executing Cluster Transition: * Resource action: FAKECLONE2:4 monitor=60000 on remote1 * Resource action: FAKECLONE2:5 monitor=60000 on 18builder * Resource action: FAKE4 monitor=60000 on remote1 +warning: Ignoring resource-discovery-enabled attribute for 18node1 because disabling resource discovery is not allowed for cluster nodes +warning: Ignoring resource-discovery-enabled attribute for 18node2 because disabling resource discovery is not allowed for cluster nodes Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/stop-failure-no-fencing.summary b/cts/scheduler/summary/stop-failure-no-fencing.summary index bb164fd5be4..9d7cd66ff56 100644 --- a/cts/scheduler/summary/stop-failure-no-fencing.summary +++ b/cts/scheduler/summary/stop-failure-no-fencing.summary @@ -1,3 +1,6 @@ +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity 0 of 9 resource instances DISABLED and 1 BLOCKED from further action due to failure Current cluster status: @@ -10,6 +13,11 @@ Current cluster status: * Clone Set: dlm-clone [dlm]: * Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * ClusterIP (ocf:heartbeat:IPaddr2): Stopped +warning: Node pcmk-3 is unclean but cannot be fenced +warning: Node pcmk-4 is unclean but cannot be fenced +error: Resource start-up disabled since no STONITH resources have been defined +error: Either configure some or disable STONITH with the stonith-enabled option +error: NOTE: Clusters with shared data need STONITH to ensure data integrity Transition Summary: diff --git a/cts/scheduler/summary/stop-failure-no-quorum.summary b/cts/scheduler/summary/stop-failure-no-quorum.summary index e76827ddfc2..a516415c28a 100644 --- a/cts/scheduler/summary/stop-failure-no-quorum.summary +++ b/cts/scheduler/summary/stop-failure-no-quorum.summary @@ -16,6 +16,8 @@ Current cluster status: * Stopped: [ pcmk-1 pcmk-3 pcmk-4 ] * ClusterIP (ocf:heartbeat:IPaddr2): Stopped * Fencing (stonith:fence_xvm): Stopped +warning: Node pcmk-3 is unclean but cannot be fenced +warning: Node pcmk-4 is unclean but cannot be fenced Transition Summary: * Fence (reboot) pcmk-2 'clvm:0 failed there' diff --git a/cts/scheduler/summary/stop-failure-with-fencing.summary b/cts/scheduler/summary/stop-failure-with-fencing.summary index 437708ef2e2..9048b95ba6a 100644 --- a/cts/scheduler/summary/stop-failure-with-fencing.summary +++ b/cts/scheduler/summary/stop-failure-with-fencing.summary @@ -13,6 +13,8 @@ Current cluster status: * Stopped: [ pcmk-1 pcmk-3 pcmk-4 ] * ClusterIP (ocf:heartbeat:IPaddr2): Stopped * Fencing (stonith:fence_xvm): Stopped +warning: Node pcmk-3 is unclean but cannot be fenced +warning: Node pcmk-4 is unclean but cannot be fenced Transition Summary: * Fence (reboot) pcmk-2 'clvm:0 failed there' diff --git a/cts/scheduler/summary/target-1.summary b/cts/scheduler/summary/target-1.summary index edc1daf32b5..0c9572b366c 100644 --- a/cts/scheduler/summary/target-1.summary +++ b/cts/scheduler/summary/target-1.summary @@ -1,3 +1,5 @@ +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones 1 of 5 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: @@ -29,6 +31,11 @@ Executing Cluster Transition: * Resource action: rsc_c001n01 monitor on c001n08 * Resource action: rsc_c001n01 monitor on c001n03 * Resource action: rsc_c001n01 monitor on c001n02 +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/target-2.summary b/cts/scheduler/summary/target-2.summary index a6194ae01ef..c39a2aa6b29 100644 --- a/cts/scheduler/summary/target-2.summary +++ b/cts/scheduler/summary/target-2.summary @@ -1,3 +1,7 @@ +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones 1 of 5 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: @@ -31,6 +35,16 @@ Executing Cluster Transition: * Resource action: rsc_c001n01 monitor on c001n08 * Resource action: rsc_c001n01 monitor on c001n03 * Resource action: rsc_c001n01 monitor on c001n02 +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n02 because 'Unpromoted' only makes sense for promotable clones +error: Ignoring 'target-role' for rsc_c001n03 because 'Promoted' only makes sense for promotable clones Revised Cluster Status: * Node List: diff --git a/cts/scheduler/summary/template-coloc-3.summary b/cts/scheduler/summary/template-coloc-3.summary index a7ff63e8dec..b26ffea9b19 100644 --- a/cts/scheduler/summary/template-coloc-3.summary +++ b/cts/scheduler/summary/template-coloc-3.summary @@ -9,6 +9,7 @@ Current cluster status: * rsc4 (ocf:pacemaker:Dummy): Stopped * rsc5 (ocf:pacemaker:Dummy): Stopped * rsc6 (ocf:pacemaker:Dummy): Stopped +error: Ignoring constraint 'template1-colo-template2' because two templates or tags cannot be colocated Transition Summary: * Start rsc1 ( node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-1.summary b/cts/scheduler/summary/ticket-promoted-1.summary index 6bc13645dfb..5bd56c510a1 100644 --- a/cts/scheduler/summary/ticket-promoted-1.summary +++ b/cts/scheduler/summary/ticket-promoted-1.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-10.summary b/cts/scheduler/summary/ticket-promoted-10.summary index eab3d91008b..c9133fe985b 100644 --- a/cts/scheduler/summary/ticket-promoted-10.summary +++ b/cts/scheduler/summary/ticket-promoted-10.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-promoted-11.summary b/cts/scheduler/summary/ticket-promoted-11.summary index 381603997eb..9bd1f55eb90 100644 --- a/cts/scheduler/summary/ticket-promoted-11.summary +++ b/cts/scheduler/summary/ticket-promoted-11.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Promote rsc1:0 ( Unpromoted -> Promoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-12.summary b/cts/scheduler/summary/ticket-promoted-12.summary index b51c277faf7..68768df73bb 100644 --- a/cts/scheduler/summary/ticket-promoted-12.summary +++ b/cts/scheduler/summary/ticket-promoted-12.summary @@ -7,6 +7,10 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-13.summary b/cts/scheduler/summary/ticket-promoted-13.summary index 6b5d14a64dd..821da141782 100644 --- a/cts/scheduler/summary/ticket-promoted-13.summary +++ b/cts/scheduler/summary/ticket-promoted-13.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-14.summary b/cts/scheduler/summary/ticket-promoted-14.summary index ee8912b2e97..31c16b5b4de 100644 --- a/cts/scheduler/summary/ticket-promoted-14.summary +++ b/cts/scheduler/summary/ticket-promoted-14.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1:0 ( Promoted node1 ) due to node availability diff --git a/cts/scheduler/summary/ticket-promoted-15.summary b/cts/scheduler/summary/ticket-promoted-15.summary index ee8912b2e97..31c16b5b4de 100644 --- a/cts/scheduler/summary/ticket-promoted-15.summary +++ b/cts/scheduler/summary/ticket-promoted-15.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1:0 ( Promoted node1 ) due to node availability diff --git a/cts/scheduler/summary/ticket-promoted-16.summary b/cts/scheduler/summary/ticket-promoted-16.summary index 851e54ebd50..a71fb4a7f8f 100644 --- a/cts/scheduler/summary/ticket-promoted-16.summary +++ b/cts/scheduler/summary/ticket-promoted-16.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-17.summary b/cts/scheduler/summary/ticket-promoted-17.summary index ee25f92c4e2..3ff57a331ec 100644 --- a/cts/scheduler/summary/ticket-promoted-17.summary +++ b/cts/scheduler/summary/ticket-promoted-17.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Demote rsc1:0 ( Promoted -> Unpromoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-18.summary b/cts/scheduler/summary/ticket-promoted-18.summary index ee25f92c4e2..3ff57a331ec 100644 --- a/cts/scheduler/summary/ticket-promoted-18.summary +++ b/cts/scheduler/summary/ticket-promoted-18.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Demote rsc1:0 ( Promoted -> Unpromoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-19.summary b/cts/scheduler/summary/ticket-promoted-19.summary index 851e54ebd50..a71fb4a7f8f 100644 --- a/cts/scheduler/summary/ticket-promoted-19.summary +++ b/cts/scheduler/summary/ticket-promoted-19.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-2.summary b/cts/scheduler/summary/ticket-promoted-2.summary index dc67f96156b..1c5370a6800 100644 --- a/cts/scheduler/summary/ticket-promoted-2.summary +++ b/cts/scheduler/summary/ticket-promoted-2.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-promoted-20.summary b/cts/scheduler/summary/ticket-promoted-20.summary index ee25f92c4e2..3ff57a331ec 100644 --- a/cts/scheduler/summary/ticket-promoted-20.summary +++ b/cts/scheduler/summary/ticket-promoted-20.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Demote rsc1:0 ( Promoted -> Unpromoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-21.summary b/cts/scheduler/summary/ticket-promoted-21.summary index f116a2eea0b..c4b3a55fb41 100644 --- a/cts/scheduler/summary/ticket-promoted-21.summary +++ b/cts/scheduler/summary/ticket-promoted-21.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Fence (reboot) node1 'deadman ticket was lost' diff --git a/cts/scheduler/summary/ticket-promoted-22.summary b/cts/scheduler/summary/ticket-promoted-22.summary index 851e54ebd50..a71fb4a7f8f 100644 --- a/cts/scheduler/summary/ticket-promoted-22.summary +++ b/cts/scheduler/summary/ticket-promoted-22.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-23.summary b/cts/scheduler/summary/ticket-promoted-23.summary index ee25f92c4e2..3ff57a331ec 100644 --- a/cts/scheduler/summary/ticket-promoted-23.summary +++ b/cts/scheduler/summary/ticket-promoted-23.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Demote rsc1:0 ( Promoted -> Unpromoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-24.summary b/cts/scheduler/summary/ticket-promoted-24.summary index b51c277faf7..68768df73bb 100644 --- a/cts/scheduler/summary/ticket-promoted-24.summary +++ b/cts/scheduler/summary/ticket-promoted-24.summary @@ -7,6 +7,10 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-promoted-3.summary b/cts/scheduler/summary/ticket-promoted-3.summary index ee8912b2e97..31c16b5b4de 100644 --- a/cts/scheduler/summary/ticket-promoted-3.summary +++ b/cts/scheduler/summary/ticket-promoted-3.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1:0 ( Promoted node1 ) due to node availability diff --git a/cts/scheduler/summary/ticket-promoted-4.summary b/cts/scheduler/summary/ticket-promoted-4.summary index eab3d91008b..c9133fe985b 100644 --- a/cts/scheduler/summary/ticket-promoted-4.summary +++ b/cts/scheduler/summary/ticket-promoted-4.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-promoted-5.summary b/cts/scheduler/summary/ticket-promoted-5.summary index 381603997eb..9bd1f55eb90 100644 --- a/cts/scheduler/summary/ticket-promoted-5.summary +++ b/cts/scheduler/summary/ticket-promoted-5.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Promote rsc1:0 ( Unpromoted -> Promoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-6.summary b/cts/scheduler/summary/ticket-promoted-6.summary index ee25f92c4e2..3ff57a331ec 100644 --- a/cts/scheduler/summary/ticket-promoted-6.summary +++ b/cts/scheduler/summary/ticket-promoted-6.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Demote rsc1:0 ( Promoted -> Unpromoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-7.summary b/cts/scheduler/summary/ticket-promoted-7.summary index eab3d91008b..c9133fe985b 100644 --- a/cts/scheduler/summary/ticket-promoted-7.summary +++ b/cts/scheduler/summary/ticket-promoted-7.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-promoted-8.summary b/cts/scheduler/summary/ticket-promoted-8.summary index 381603997eb..9bd1f55eb90 100644 --- a/cts/scheduler/summary/ticket-promoted-8.summary +++ b/cts/scheduler/summary/ticket-promoted-8.summary @@ -6,6 +6,14 @@ Current cluster status: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Promote rsc1:0 ( Unpromoted -> Promoted node1 ) diff --git a/cts/scheduler/summary/ticket-promoted-9.summary b/cts/scheduler/summary/ticket-promoted-9.summary index f116a2eea0b..c4b3a55fb41 100644 --- a/cts/scheduler/summary/ticket-promoted-9.summary +++ b/cts/scheduler/summary/ticket-promoted-9.summary @@ -7,6 +7,14 @@ Current cluster status: * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc1-monitor-unpromoted-5 is duplicate of rsc1-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Fence (reboot) node1 'deadman ticket was lost' diff --git a/cts/scheduler/summary/ticket-rsc-sets-1.summary b/cts/scheduler/summary/ticket-rsc-sets-1.summary index d119ce5176e..e7a300c5a2c 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-1.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-1.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc5:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-rsc-sets-10.summary b/cts/scheduler/summary/ticket-rsc-sets-10.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-10.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-10.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-11.summary b/cts/scheduler/summary/ticket-rsc-sets-11.summary index 03153aa264b..2775ac69305 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-11.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-11.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-rsc-sets-12.summary b/cts/scheduler/summary/ticket-rsc-sets-12.summary index 68e0827f78b..b387a94fcd4 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-12.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-12.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-13.summary b/cts/scheduler/summary/ticket-rsc-sets-13.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-13.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-13.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-14.summary b/cts/scheduler/summary/ticket-rsc-sets-14.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-14.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-14.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-2.summary b/cts/scheduler/summary/ticket-rsc-sets-2.summary index fccf3cad1ba..5e6c47b66f7 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-2.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-2.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1 ( node2 ) diff --git a/cts/scheduler/summary/ticket-rsc-sets-3.summary b/cts/scheduler/summary/ticket-rsc-sets-3.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-3.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-3.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-4.summary b/cts/scheduler/summary/ticket-rsc-sets-4.summary index d119ce5176e..e7a300c5a2c 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-4.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-4.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Stopped: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc5:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-rsc-sets-5.summary b/cts/scheduler/summary/ticket-rsc-sets-5.summary index 217243a7b25..9d808a2ebd8 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-5.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-5.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc1 ( node2 ) diff --git a/cts/scheduler/summary/ticket-rsc-sets-6.summary b/cts/scheduler/summary/ticket-rsc-sets-6.summary index 7336f70db30..4d446693ea9 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-6.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-6.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Start rsc4:0 ( node2 ) diff --git a/cts/scheduler/summary/ticket-rsc-sets-7.summary b/cts/scheduler/summary/ticket-rsc-sets-7.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-7.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-7.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/ticket-rsc-sets-8.summary b/cts/scheduler/summary/ticket-rsc-sets-8.summary index 03153aa264b..2775ac69305 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-8.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-8.summary @@ -12,6 +12,14 @@ Current cluster status: * Stopped: [ node1 node2 ] * Clone Set: ms5 [rsc5] (promotable): * Unpromoted: [ node1 node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: diff --git a/cts/scheduler/summary/ticket-rsc-sets-9.summary b/cts/scheduler/summary/ticket-rsc-sets-9.summary index 3bc9d648ac3..f8612ba8a2f 100644 --- a/cts/scheduler/summary/ticket-rsc-sets-9.summary +++ b/cts/scheduler/summary/ticket-rsc-sets-9.summary @@ -13,6 +13,14 @@ Current cluster status: * Clone Set: ms5 [rsc5] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) +error: Operation rsc5-monitor-unpromoted-5 is duplicate of rsc5-monitor-promoted-5 (do not use same name and interval combination more than once per resource) Transition Summary: * Stop rsc1 ( node2 ) due to node availability diff --git a/cts/scheduler/summary/unrunnable-1.summary b/cts/scheduler/summary/unrunnable-1.summary index 75fda238563..9ba6f2ecf54 100644 --- a/cts/scheduler/summary/unrunnable-1.summary +++ b/cts/scheduler/summary/unrunnable-1.summary @@ -18,6 +18,7 @@ Current cluster status: * child_DoFencing:1 (stonith:ssh): Started c001n02 (UNCLEAN) * child_DoFencing:2 (stonith:ssh): Stopped * child_DoFencing:3 (stonith:ssh): Stopped +warning: Node c001n02 is unclean but cannot be fenced Transition Summary: * Start DcIPaddr ( c001n03 ) due to no quorum (blocked) diff --git a/cts/scheduler/summary/unrunnable-2.summary b/cts/scheduler/summary/unrunnable-2.summary index 26c63510785..0c0ee882ad4 100644 --- a/cts/scheduler/summary/unrunnable-2.summary +++ b/cts/scheduler/summary/unrunnable-2.summary @@ -85,6 +85,7 @@ Current cluster status: * Stopped: [ overcloud-controller-0 overcloud-controller-1 overcloud-controller-2 ] * Clone Set: openstack-nova-conductor-clone [openstack-nova-conductor]: * Stopped: [ overcloud-controller-0 overcloud-controller-1 overcloud-controller-2 ] +warning: Support for require-all in ordering constraints is deprecated and will be removed in a future release (use clone-min clone meta-attribute instead) Transition Summary: * Start openstack-cinder-volume ( overcloud-controller-2 ) due to unrunnable openstack-cinder-scheduler-clone running (blocked) diff --git a/cts/scheduler/summary/whitebox-imply-stop-on-fence.summary b/cts/scheduler/summary/whitebox-imply-stop-on-fence.summary index 78506c5354e..79c058252dd 100644 --- a/cts/scheduler/summary/whitebox-imply-stop-on-fence.summary +++ b/cts/scheduler/summary/whitebox-imply-stop-on-fence.summary @@ -24,6 +24,12 @@ Current cluster status: * R-lxc-01_kiff-02 (ocf:heartbeat:VirtualDomain): Started kiff-02 * R-lxc-02_kiff-02 (ocf:heartbeat:VirtualDomain): Started kiff-02 * vm-fs (ocf:heartbeat:Filesystem): FAILED lxc-01_kiff-01 +warning: Invalid ordering constraint between shared0:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between clvmd:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between dlm:0 and R-lxc-02_kiff-02 +warning: Invalid ordering constraint between shared0:0 and R-lxc-01_kiff-02 +warning: Invalid ordering constraint between clvmd:0 and R-lxc-01_kiff-02 +warning: Invalid ordering constraint between dlm:0 and R-lxc-01_kiff-02 Transition Summary: * Fence (reboot) lxc-02_kiff-01 (resource: R-lxc-02_kiff-01) 'guest is unclean'