Sign Up
Log In
Log In
or
Sign Up
Places
All Projects
Status Monitor
Collapse sidebar
openSUSE:Step:15
pacemaker.19404
bsc#1177212-0008-Test-scheduler-test-failed-mig...
Overview
Repositories
Revisions
Requests
Users
Attributes
Meta
File bsc#1177212-0008-Test-scheduler-test-failed-migration-followed-by-suc.patch of Package pacemaker.19404
From 64834f3ac99e8f8286397bda172f21f0e1cb2e75 Mon Sep 17 00:00:00 2001 From: Ken Gaillot <kgaillot@redhat.com> Date: Fri, 5 Jul 2019 19:34:05 -0500 Subject: [PATCH 8/8] Test: scheduler: test failed migration followed by successful reverse migration --- cts/cts-pengine.in | 1 + cts/pengine/migration-ping-pong.dot | 2 + cts/pengine/migration-ping-pong.exp | 1 + cts/pengine/migration-ping-pong.scores | 10 ++ cts/pengine/migration-ping-pong.summary | 25 ++++ cts/pengine/migration-ping-pong.xml | 165 ++++++++++++++++++++++ 6 files changed, 204 insertions(+) create mode 100644 cts/pengine/migration-ping-pong.dot create mode 100644 cts/pengine/migration-ping-pong.exp create mode 100644 cts/pengine/migration-ping-pong.scores create mode 100644 cts/pengine/migration-ping-pong.summary create mode 100644 cts/pengine/migration-ping-pong.xml Index: pacemaker-1.1.18+20180430.b12c320f5/cts/cts-pengine.in =================================================================== --- pacemaker-1.1.18+20180430.b12c320f5.orig/cts/cts-pengine.in +++ pacemaker-1.1.18+20180430.b12c320f5/cts/cts-pengine.in @@ -653,6 +653,7 @@ do_test migrate-fail-6 "Failed migrate_t do_test migrate-fail-7 "Failed migrate_to + stop on source" do_test migrate-fail-8 "Failed migrate_to + stop on target - ideally we wouldn't need to re-stop on target" do_test migrate-fail-9 "Failed migrate_to + stop on source and target" +do_test migration-ping-pong "Old migrate_to failure + successful migrate_from on same node" do_test migrate-stop "Migration in a stopping stack" do_test migrate-start "Migration in a starting stack" Index: pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.dot =================================================================== --- /dev/null +++ pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.dot @@ -0,0 +1,2 @@ + digraph "g" { +} Index: pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.exp =================================================================== --- /dev/null +++ pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.exp @@ -0,0 +1 @@ +<transition_graph cluster-delay="60s" stonith-timeout="60s" failed-stop-offset="INFINITY" failed-start-offset="INFINITY" transition_id="0"/> Index: pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.scores =================================================================== --- /dev/null +++ pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.scores @@ -0,0 +1,10 @@ +Allocation scores: +Using the original execution date of: 2019-06-06 13:56:45Z +native_color: fence_ilo_ha-idg-1 allocation score on ha-idg-1: -INFINITY +native_color: fence_ilo_ha-idg-1 allocation score on ha-idg-2: 0 +native_color: fence_ilo_ha-idg-2 allocation score on ha-idg-1: 200 +native_color: fence_ilo_ha-idg-2 allocation score on ha-idg-2: -INFINITY +native_color: vm_idcc_devel allocation score on ha-idg-1: 200 +native_color: vm_idcc_devel allocation score on ha-idg-2: 0 +native_color: vm_severin allocation score on ha-idg-1: 200 +native_color: vm_severin allocation score on ha-idg-2: 0 Index: pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.summary =================================================================== --- /dev/null +++ pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.summary @@ -0,0 +1,25 @@ +Using the original execution date of: 2019-06-06 13:56:45Z + +Current cluster status: +Node ha-idg-2 (1084777492): standby +Online: [ ha-idg-1 ] + + fence_ilo_ha-idg-2 (stonith:fence_ilo2): Started ha-idg-1 + fence_ilo_ha-idg-1 (stonith:fence_ilo4): Stopped + vm_idcc_devel (ocf::heartbeat:VirtualDomain): Started ha-idg-1 + vm_severin (ocf::heartbeat:VirtualDomain): Started ha-idg-1 + +Transition Summary: + +Executing cluster transition: +Using the original execution date of: 2019-06-06 13:56:45Z + +Revised cluster status: +Node ha-idg-2 (1084777492): standby +Online: [ ha-idg-1 ] + + fence_ilo_ha-idg-2 (stonith:fence_ilo2): Started ha-idg-1 + fence_ilo_ha-idg-1 (stonith:fence_ilo4): Stopped + vm_idcc_devel (ocf::heartbeat:VirtualDomain): Started ha-idg-1 + vm_severin (ocf::heartbeat:VirtualDomain): Started ha-idg-1 + Index: pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.xml =================================================================== --- /dev/null +++ pacemaker-1.1.18+20180430.b12c320f5/cts/pengine/migration-ping-pong.xml @@ -0,0 +1,165 @@ +<cib crm_feature_set="3.0.14" validate-with="pacemaker-2.7" epoch="6815" num_updates="15" admin_epoch="2" cib-last-written="Thu Jun 6 15:41:29 2019" update-origin="ha-idg-2" update-client="crm_attribute" update-user="root" have-quorum="1" dc-uuid="1084777482" execution-date="1559829405"> + <configuration> + <crm_config> + <cluster_property_set id="cib-bootstrap-options"> + <nvpair name="have-watchdog" value="false" id="cib-bootstrap-options-have-watchdog"/> + <nvpair name="dc-version" value="1.1.19+20181105.ccd6b5b10-3.3.1-1.1.19+20181105.ccd6b5b10" id="cib-bootstrap-options-dc-version"/> + <nvpair name="cluster-infrastructure" value="corosync" id="cib-bootstrap-options-cluster-infrastructure"/> + <nvpair name="last-lrm-refresh" value="1559146053" id="cib-bootstrap-options-last-lrm-refresh"/> + <nvpair name="cluster-name" value="ha-idg" id="cib-bootstrap-options-cluster-name"/> + <nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/> + <nvpair name="stonith-enabled" value="true" id="cib-bootstrap-options-stonith-enabled"/> + <nvpair name="stonith-action" value="Off" id="cib-bootstrap-options-stonith-action"/> + </cluster_property_set> + </crm_config> + <nodes> + <node id="1084777492" uname="ha-idg-2"> + <instance_attributes id="nodes-1084777492"> + <nvpair id="nodes-1084777492-maintenance" name="maintenance" value="off"/> + <nvpair id="nodes-1084777492-standby" name="standby" value="on"/> + </instance_attributes> + </node> + <node id="1084777482" uname="ha-idg-1"> + <instance_attributes id="nodes-1084777482"> + <nvpair id="nodes-1084777482-maintenance" name="maintenance" value="off"/> + <nvpair id="nodes-1084777482-standby" name="standby" value="off"/> + </instance_attributes> + </node> + </nodes> + <resources> + <primitive id="fence_ilo_ha-idg-2" class="stonith" type="fence_ilo2" description="fenct ha-idg-2 mit ILO"> + <operations> + <op name="monitor" interval="30m" timeout="120s" id="fence_ha-idg-2-monitor-30m"/> + </operations> + </primitive> + <primitive id="fence_ilo_ha-idg-1" class="stonith" type="fence_ilo4" description="fenct ha-idg-1 mit ILO"> + <operations> + <op name="monitor" interval="30m" timeout="120s" id="fence_ha-idg-1-monitor-30m"/> + </operations> + </primitive> + <primitive id="vm_idcc_devel" class="ocf" provider="heartbeat" type="VirtualDomain"> + <instance_attributes id="vm_idcc-devel-instance_attributes"> + <nvpair name="config" value="/mnt/share/idcc_devel.xml" id="vm_idcc-devel-instance_attributes-config"/> + </instance_attributes> + <instance_attributes id="vm_idcc-devel-instance_attributes-0"> + <nvpair name="hypervisor" value="qemu:///system" id="vm_idcc-devel-instance_attributes-0-hypervisor"/> + </instance_attributes> + <instance_attributes id="vm_idcc-devel-instance_attributes-1"> + <nvpair name="migration_transport" value="ssh" id="vm_idcc-devel-instance_attributes-1-migration_transport"/> + </instance_attributes> + <instance_attributes id="vm_idcc-devel-instance_attributes-2"> + <nvpair name="migration_network_suffix" value="-private" id="vm_idcc-devel-instance_attributes-2-migration_network_suffix"/> + </instance_attributes> + <operations> + <op name="start" interval="0" timeout="120" id="vm_idcc-devel-start-0"/> + <op name="stop" interval="0" timeout="130" id="vm_idcc-devel-stop-0"/> + <op name="monitor" interval="30" timeout="25" id="vm_idcc-devel-monitor-30"/> + <op name="migrate_from" interval="0" timeout="300" id="vm_idcc-devel-migrate_from-0"/> + <op name="migrate_to" interval="0" timeout="300" id="vm_idcc-devel-migrate_to-0"/> + </operations> + <meta_attributes id="vm_idcc-devel-meta_attributes"> + <nvpair name="allow-migrate" value="true" id="vm_idcc-devel-meta_attributes-allow-migrate"/> + <nvpair name="target-role" value="Started" id="vm_idcc-devel-meta_attributes-target-role"/> + <nvpair name="is-managed" value="true" id="vm_idcc-devel-meta_attributes-is-managed"/> + </meta_attributes> + <utilization id="vm_idcc-devel-utilization"> + <nvpair name="cpu" value="1" id="vm_idcc-devel-utilization-cpu"/> + <nvpair name="hv_memory" value="1020" id="vm_idcc-devel-utilization-hv_memory"/> + </utilization> + </primitive> + <primitive id="vm_severin" class="ocf" provider="heartbeat" type="VirtualDomain"> + <instance_attributes id="vm_severin-instance_attributes"> + <nvpair name="config" value="/mnt/share/severin.xml" id="vm_severin-instance_attributes-config"/> + </instance_attributes> + <instance_attributes id="vm_severin-instance_attributes-0"> + <nvpair name="hypervisor" value="qemu:///system" id="vm_severin-instance_attributes-0-hypervisor"/> + </instance_attributes> + <instance_attributes id="vm_severin-instance_attributes-1"> + <nvpair name="migration_transport" value="ssh" id="vm_severin-instance_attributes-1-migration_transport"/> + </instance_attributes> + <instance_attributes id="vm_severin-instance_attributes-2"> + <nvpair name="migration_network_suffix" value="-private" id="vm_severin-instance_attributes-2-migration_network_suffix"/> + </instance_attributes> + <operations> + <op name="start" interval="0" timeout="120" id="vm_severin-start-0"/> + <op name="stop" interval="0" timeout="130" id="vm_severin-stop-0"/> + <op name="monitor" interval="30" timeout="25" id="vm_severin-monitor-30"/> + <op name="migrate_from" interval="0" timeout="300" id="vm_severin-migrate_from-0"/> + <op name="migrate_to" interval="0" timeout="300" id="vm_severin-migrate_to-0"/> + </operations> + <meta_attributes id="vm_severin-meta_attributes"> + <nvpair name="allow-migrate" value="true" id="vm_severin-meta_attributes-allow-migrate"/> + <nvpair name="target-role" value="Started" id="vm_severin-meta_attributes-target-role"/> + <nvpair name="is-managed" value="true" id="vm_severin-meta_attributes-is-managed"/> + </meta_attributes> + <utilization id="vm_severin-utilization"> + <nvpair name="cpu" value="2" id="vm_severin-utilization-cpu"/> + <nvpair name="hv_memory" value="2176" id="vm_severin-utilization-hv_memory"/> + </utilization> + </primitive> + </resources> + <constraints> + <rsc_location id="loc_fence_ilo_ha-idg-2" rsc="fence_ilo_ha-idg-2" score="-INFINITY" node="ha-idg-2"/> + <rsc_location id="loc_fence_ilo_ha-idg-1" rsc="fence_ilo_ha-idg-1" score="-INFINITY" node="ha-idg-1"/> + </constraints> + <rsc_defaults> + <meta_attributes id="rsc-options"> + <nvpair name="resource-stickiness" value="200" id="rsc-options-resource-stickiness"/> + </meta_attributes> + </rsc_defaults> + <alerts/> + </configuration> + <status> + <node_state id="1084777482" uname="ha-idg-1" in_ccm="true" crmd="online" crm-debug-origin="do_update_resource" join="member" expected="member"> + <transient_attributes id="1084777482"> + <instance_attributes id="status-1084777482"> + <nvpair id="status-1084777482-fail-count-vm_idcc_devel.monitor_30000" name="fail-count-vm_idcc_devel#monitor_30000" value="4"/> + <nvpair id="status-1084777482-last-failure-vm_idcc_devel.monitor_30000" name="last-failure-vm_idcc_devel#monitor_30000" value="1559766644"/> + </instance_attributes> + </transient_attributes> + <lrm id="1084777482"> + <lrm_resources> + <lrm_resource id="fence_ilo_ha-idg-1" type="fence_ilo4" class="stonith"> + <lrm_rsc_op id="fence_ilo_ha-idg-1_last_0" operation_key="fence_ilo_ha-idg-1_monitor_0" operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.14" transition-key="3:0:7:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:7;3:0:7:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-1" call-id="9" rc-code="7" op-status="0" interval="0" last-run="1558969612" last-rc-change="1558969612" exec-time="0" queue-time="0" op-digest="3954a01b79da87e4607bb152bdb8fc12" op-secure-params=" password passwd " op-secure-digest="ba60b2919b6ff526e7844689cbedee60"/> + </lrm_resource> + <lrm_resource id="fence_ilo_ha-idg-2" type="fence_ilo2" class="stonith"> + <lrm_rsc_op id="fence_ilo_ha-idg-2_last_0" operation_key="fence_ilo_ha-idg-2_start_0" operation="start" crm-debug-origin="crm_simulate" crm_feature_set="3.2.0" transition-key="279:-1:0:xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" transition-magic="0:0;279:-1:0:xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" exit-reason="" on_node="ha-idg-1" call-id="279" rc-code="0" op-status="0" interval="0" last-run="1562373136" last-rc-change="1562373136" exec-time="0" queue-time="0" op-digest="f2317cad3d54cec5d7d7aa7d0bf35cf8" op-secure-params=" password passwd " op-secure-digest="bddd7b99d0ff68fb24869b2682af58e1"/> + <lrm_rsc_op id="fence_ilo_ha-idg-2_monitor_1800000" operation_key="fence_ilo_ha-idg-2_monitor_1800000" operation="monitor" crm-debug-origin="crm_simulate" crm_feature_set="3.2.0" transition-key="280:-1:0:xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" transition-magic="0:0;280:-1:0:xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx" exit-reason="" on_node="ha-idg-1" call-id="280" rc-code="0" op-status="0" interval="1800000" last-rc-change="1562373136" exec-time="0" queue-time="0" op-digest="0e121d086272acfe9236910e5fe2e8cd" op-secure-params=" password passwd " op-secure-digest="bddd7b99d0ff68fb24869b2682af58e1"/> + </lrm_resource> + <lrm_resource id="vm_idcc_devel" type="VirtualDomain" class="ocf" provider="heartbeat"> + <lrm_rsc_op id="vm_idcc_devel_last_failure_0" operation_key="vm_idcc_devel_migrate_to_0" operation="migrate_to" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="57:1178:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:1;57:1178:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="idcc_devel: live migration to ha-idg-2-private failed: 1" on_node="ha-idg-1" call-id="427" rc-code="1" op-status="0" interval="0" last-rc-change="1559811643" exec-time="657" queue-time="4" op-digest="006e1cb46abea81d42ca1f06bdcf2a15" last-run="1559811643" migrate_source="ha-idg-1" migrate_target="ha-idg-2"/> + <lrm_rsc_op id="vm_idcc_devel_last_0" operation_key="vm_idcc_devel_migrate_from_0" operation="migrate_from" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="57:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;57:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-1" call-id="432" rc-code="0" op-status="0" interval="0" last-run="1559828500" last-rc-change="1559828500" exec-time="1958" queue-time="0" op-digest="006e1cb46abea81d42ca1f06bdcf2a15" migrate_source="ha-idg-2" migrate_target="ha-idg-1"/> + <lrm_rsc_op id="vm_idcc_devel_monitor_30000" operation_key="vm_idcc_devel_monitor_30000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="55:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;55:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-1" call-id="436" rc-code="0" op-status="0" interval="30000" last-rc-change="1559828502" exec-time="199" queue-time="0" op-digest="89602f284e214dd8243cc82b08200ead"/> + </lrm_resource> + <lrm_resource id="vm_severin" type="VirtualDomain" class="ocf" provider="heartbeat"> + <lrm_rsc_op id="vm_severin_last_failure_0" operation_key="vm_severin_migrate_to_0" operation="migrate_to" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.14" transition-key="85:273:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:1;85:273:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="severin: live migration to ha-idg-2-private failed: 1" on_node="ha-idg-1" call-id="219" rc-code="1" op-status="0" interval="0" last-run="1559144036" last-rc-change="1559144036" exec-time="35560" queue-time="0" migrate_source="ha-idg-1" migrate_target="ha-idg-2" op-digest="76cbfd60c4831f62e13e997133576aa2"/> + <lrm_rsc_op id="vm_severin_last_0" operation_key="vm_severin_migrate_from_0" operation="migrate_from" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="68:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;68:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-1" call-id="433" rc-code="0" op-status="0" interval="0" last-run="1559828500" last-rc-change="1559828500" exec-time="1964" queue-time="0" migrate_source="ha-idg-2" migrate_target="ha-idg-1" op-digest="76cbfd60c4831f62e13e997133576aa2"/> + <lrm_rsc_op id="vm_severin_monitor_30000" operation_key="vm_severin_monitor_30000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="66:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;66:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-1" call-id="437" rc-code="0" op-status="0" interval="30000" last-rc-change="1559828502" exec-time="199" queue-time="0" op-digest="6f1fd783ab2d52e9fb1565c166bf12a3"/> + </lrm_resource> + </lrm_resources> + </lrm> + </node_state> + <node_state id="1084777492" uname="ha-idg-2" crmd="online" crm-debug-origin="do_update_resource" in_ccm="true" join="member" expected="member"> + <lrm id="1084777492"> + <lrm_resources> + <lrm_resource id="fence_ilo_ha-idg-2" type="fence_ilo2" class="stonith"> + <lrm_rsc_op id="fence_ilo_ha-idg-2_last_0" operation_key="fence_ilo_ha-idg-2_monitor_0" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="15:1177:7:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:7;15:1177:7:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="5" rc-code="7" op-status="0" interval="0" last-run="1559811561" last-rc-change="1559811561" exec-time="1" queue-time="1" op-digest="d065315be62598c16e3f435f4add0e40" op-secure-params=" password passwd " op-secure-digest="bddd7b99d0ff68fb24869b2682af58e1"/> + </lrm_resource> + <lrm_resource id="fence_ilo_ha-idg-1" type="fence_ilo4" class="stonith"> + <lrm_rsc_op id="fence_ilo_ha-idg-1_last_0" operation_key="fence_ilo_ha-idg-1_stop_0" operation="stop" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="22:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;22:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="96" rc-code="0" op-status="0" interval="0" last-run="1559828489" last-rc-change="1559828489" exec-time="1" queue-time="0" op-digest="3954a01b79da87e4607bb152bdb8fc12" op-secure-params=" password passwd " op-secure-digest="ba60b2919b6ff526e7844689cbedee60"/> + <lrm_rsc_op id="fence_ilo_ha-idg-1_monitor_1800000" operation_key="fence_ilo_ha-idg-1_monitor_1800000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="18:1178:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;18:1178:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="69" rc-code="0" op-status="0" interval="1800000" last-rc-change="1559811638" exec-time="219" queue-time="0" op-digest="714126349f6efb14dbfd462f2b7c1158" op-secure-params=" password passwd " op-secure-digest="ba60b2919b6ff526e7844689cbedee60"/> + </lrm_resource> + <lrm_resource id="vm_idcc_devel" type="VirtualDomain" class="ocf" provider="heartbeat"> + <lrm_rsc_op id="vm_idcc_devel_last_0" operation_key="vm_idcc_devel_stop_0" operation="stop" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="53:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;53:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="107" rc-code="0" op-status="0" interval="0" last-run="1559828502" last-rc-change="1559828502" exec-time="63" queue-time="0" op-digest="006e1cb46abea81d42ca1f06bdcf2a15" migrate_source="ha-idg-2" migrate_target="ha-idg-1"/> + <lrm_rsc_op id="vm_idcc_devel_monitor_30000" operation_key="vm_idcc_devel_monitor_30000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="60:1179:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;60:1179:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="93" rc-code="0" op-status="0" interval="30000" last-rc-change="1559811666" exec-time="173" queue-time="0" op-digest="89602f284e214dd8243cc82b08200ead"/> + </lrm_resource> + <lrm_resource id="vm_severin" type="VirtualDomain" class="ocf" provider="heartbeat"> + <lrm_rsc_op id="vm_severin_last_0" operation_key="vm_severin_stop_0" operation="stop" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="64:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;64:1199:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="108" rc-code="0" op-status="0" interval="0" last-run="1559828502" last-rc-change="1559828502" exec-time="60" queue-time="0" op-digest="76cbfd60c4831f62e13e997133576aa2" migrate_source="ha-idg-2" migrate_target="ha-idg-1"/> + <lrm_rsc_op id="vm_severin_last_failure_0" operation_key="vm_severin_monitor_0" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="28:1177:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:7;28:1177:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="62" rc-code="7" op-status="0" interval="0" last-run="1559811563" last-rc-change="1559811563" exec-time="47" queue-time="105" op-digest="76cbfd60c4831f62e13e997133576aa2"/> + <lrm_rsc_op id="vm_severin_monitor_30000" operation_key="vm_severin_monitor_30000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.14" transition-key="69:1179:0:6c7d62a8-ea9a-4082-bc13-489329277d66" transition-magic="0:0;69:1179:0:6c7d62a8-ea9a-4082-bc13-489329277d66" exit-reason="" on_node="ha-idg-2" call-id="88" rc-code="0" op-status="0" interval="30000" last-rc-change="1559811644" exec-time="185" queue-time="0" op-digest="6f1fd783ab2d52e9fb1565c166bf12a3"/> + </lrm_resource> + </lrm_resources> + </lrm> + </node_state> + </status> +</cib>
Locations
Projects
Search
Status Monitor
Help
OpenBuildService.org
Documentation
API Documentation
Code of Conduct
Contact
Support
@OBShq
Terms
openSUSE Build Service is sponsored by
The Open Build Service is an
openSUSE project
.
Sign Up
Log In
Places
Places
All Projects
Status Monitor