Hi all,
I'm kinda new to Pacemaker and have some questions I hope some of you could
help me with, any idea(s) would be highly appreciated, thank you.
I'am building the HA cluster with corosync2.3.3 + cluster-glue +
pacemaker-1.1.11-rc5. The problem is, when I try to remove a node from HA
cluster with "crm_node -f -R nodeID", it seems to remove the wrong node.
My setup is as follows(and it can be reproduced):
1. The HA is ok with 2 resource and 2 nodes, and with stonith-enabled="false":
root@h1:/opt/bin# crm status
Last updated: Sat Apr 12 16:00:51 2014
Last change: Sat Apr 12 16:00:43 2014 via cibadmin on h1
Stack: corosync
Current DC: h1 (1084752017) - partition with quorum
Version: 1.1.10-33f9d09
2 Nodes configured
2 Resources configured
Online: [ h0 h1 ]
VMdos-1 (ocf::heartbeat:VirtualDomain): Started h1
root@h1:/opt/bin#
root@h1:/opt/bin# crm_node --list
1084752017 h1
1084752016 h0
root@h1:/opt/bin#
2. remove node h0 from HA
(1). Stop pacemakerd and corosync service on node h0 (2). Run the follow cmd on
node h1 to Remove node info from cib:
cibadmin --delete --obj_type status --crm_xml "<node_state id=\"1084752016\"/>"
cibadmin --delete --obj_type nodes --crm_xml "<node id=\"1084752016\"/>"
at this time; the result of "crm status" shows that node " h0 " is offline and
there's no information of node h0 in CIB. But the "crm_node --list " still
include node" h0"
3. crm_node -f -R 1084752016
After this cmd, node " h1" is removed, and result of "crm status " is as
follow(with 0 node):
root@h1:/opt/bin# crm status
Last updated: Sat Apr 12 15:59:42 2014
Last change: Sat Apr 12 15:59:37 2014 via crm_node on h1
Stack: corosync
Current DC: NONE
0 Nodes configured
2 Resources configured
root@h1:/opt/bin
The corosync.conf is as follow:
.............
quorum {
provider: corosync_votequorum
expected_votes: 2
allow_downscale: 1
two_node: 1
}
...........
The corosync.log is as follow, it seems node " h1" is removed due to fencing!
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:299 Free'ing
ringbuffer: /dev/shm/qb-cmap-request-25004-14152-27-header
Apr 12 16:22:35 [25026] h1 cib: debug: activateCibXml:
Triggering CIB write for cib_delete op
Apr 12 16:22:35 [25031] h1 crmd: notice: crm_reap_dead_member:
Removing h0/1084752016 from the membership list
Apr 12 16:22:35 [25031] h1 crmd: notice: reap_crm_member: Purged
1 peers with id=1084752016 and/or uname=(null) from the membership cache
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: log_cib_diff: Config update:
Local-only Change: 0.12.1
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <cib
admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
<configuration>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
<nodes>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<node id="1084752017" uname="h1"/>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
</nodes>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
</configuration>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
</cib>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++ <cib
epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
update-origin="h1" update-client="crm_node" have-quorum="1"
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25026] h1 cib: notice: log_cib_diff: cib:diff:
Local-only Change: 0.12.1
Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: -- <node
id="1084752017" uname="h1"/>
Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: ++ <cib
epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
update-origin="h1" update-client="crm_node" have-quorum="1"
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_delete operation for section nodes: OK (rc=0,
origin=local/crm_node/2, version=0.12.1)
Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff:
Processing diff (cib_delete): 0.11.1 -> 0.12.1 (S_IDLE)
Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph:
te_update_diff:126 - Triggered transition abort (complete=1, node=, tag=diff,
id=(null), magic=NA, cib=0.12.1) : Non-status change
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <diff crm_feature_set="3.0.8" digest="3cccef06483ac4dfeadfb562f6f8293a">
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <diff-removed admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <cib admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <configuration>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <nodes>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <node id="1084752017" uname="h1"
__crm_diff_marker__="removed:top"/>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </nodes>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </configuration>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </cib>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </diff-removed>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <diff-added>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <cib epoch="12" num_updates="1" admin_epoch="0"
validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr
12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1"
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </diff-added>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </diff>
Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
I_PE_CALC: [ state=S_IDLE cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
origin=abort_transition_graph ]
Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1
cluster nodes are eligible to run resources.
Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 38:
Requesting the current CIB: S_POLICY_ENGINE
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_query operation for section 'all': OK (rc=0,
origin=local/crmd/38, version=0.12.1)
Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback:
Invoking the PE: query=38, ref=pe_calc-dc-1397290955-21, seq=1024, quorate=1
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff:
--- 0.12.1
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff:
+++ 0.12.2 3d673c27c3c92939b41c7207edee9f46
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <cib
num_updates="1">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
<status>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<node_state id="1084752017" uname="h1" in_ccm="true" crmd="online"
crm-debug-origin="post_cache_update" join="member" expected="member">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<transient_attributes id="1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<instance_attributes id="status-1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<nvpair id="status-1084752017-probe_complete" name="probe_complete"
value="true"/>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</instance_attributes>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</transient_attributes>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm id="1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_resources>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_resource id="VMdos-1" type="VirtualDomain" class="ocf"
provider="heartbeat">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_rsc_op id="VMdos-1_last_failure_0"
operation_key="VMdos-1_monitor_0" operation="monitor"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8"
transition-key="6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c"
transition-magic="0:0;6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="5"
rc-code="0" op-status="0" interval="0" last-run="1397290607"
last-rc-change="1397290607" exec-time="92" queue-time="0" op-digest="
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_rsc_op id="VMdos-1_monitor_30000"
operation_key="VMdos-1_monitor_30000" operation="monitor"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8"
transition-key="7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c"
transition-magic="0:0;7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="6"
rc-code="0" op-status="0" interval="30000" last-rc-change="1397290607"
exec-time="67" queue-time="0" op-digest="0874c7ce5f61a12
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</lrm_resource>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_resource id="VMdos-2" type="VirtualDomain" class="ocf"
provider="heartbeat">
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
<lrm_rsc_op id="VMdos-2_last_0" operation_key="VMdos-2_monitor_0"
operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8"
transition-key="7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c"
transition-magic="0:7;7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="10"
rc-code="7" op-status="0" interval="0" last-run="1397290608"
last-rc-change="1397290608" exec-time="61" queue-time="0" op-digest="c7d22be
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</lrm_resource>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</lrm_resources>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</lrm>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
</node_state>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
</status>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
</cib>
Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++ <cib
epoch="12" num_updates="2" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
update-origin="h1" update-client="crm_node" have-quorum="1"
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH
timeout: 60000
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH
of failed nodes is disabled
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop
all active resources: false
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Cluster
is symmetric - resources can run anywhere by default
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Default
stickiness: 0
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On loss
of CCM Quorum: Stop ALL resources
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node
scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains:
Unpacking domains
Apr 12 16:22:35 [25030] h1 pengine: warning: unpack_status: Node h1
in status section no longer exists
Apr 12 16:22:35 [25030] h1 pengine: info: unpack_status: Node
1084752017 is unknown
Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
not allocate a node for VMdos-1
Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
VMdos-1 cannot run anywhere
Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
not allocate a node for VMdos-2
Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
VMdos-2 cannot run anywhere
Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-1
(Stopped)
Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-2
(Stopped)
Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message:
Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE
origin=handle_response ]
Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=handle_response ]
Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked
transition 4: 0 actions in 0 synapses
Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing
graph 4 (ref=pe_calc-dc-1397290955-21) derived from
/var/lib/pacemaker/pengine/pe-input-4.bz2
Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
transition graph
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_delete operation for section status: OK (rc=0,
origin=local/crm_node/3, version=0.12.2)
Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff:
Processing diff (cib_delete): 0.12.1 -> 0.12.2 (S_TRANSITION_ENGINE)
Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph:
te_update_diff:188 - Triggered transition abort (complete=0, node=h1,
tag=transient_attributes, id=1084752017, magic=NA, cib=0.12.2) : Transient
attribute: removal
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <transient_attributes id="1084752017">
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <instance_attributes id="status-1084752017">
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause <nvpair id="status-1084752017-probe_complete" name="probe_complete"
value="true"/>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </instance_attributes>
Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
Cause </transient_attributes>
Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority:
Abort priority upgraded from 0 to 1000000
Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority:
Abort action done superceeded by restart
Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 4
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
transition graph
Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger:
Transition 4 is now complete
Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing
transition completion in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 4
status: restart - Transient attribute: removal
Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
I_PE_CALC: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1
cluster nodes are eligible to run resources.
Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 39:
Requesting the current CIB: S_POLICY_ENGINE
Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_new:
Connecting 0x25b4ea0 for uid=0 gid=0 pid=14152
id=f3612e17-0806-4355-a3fc-2cf1feda1e6d
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: handle_new_connection:
IPC credentials authenticated (25024-14152-10)
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_shm_connect:
connecting to client [14152]
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_query operation for section 'all': OK (rc=0,
origin=local/crmd/39, version=0.12.2)
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25026] h1 cib: debug:
qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect:
qb_ipcs_disconnect(25026-14152-13) state:2
Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy:
Destroying 0 events
Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
ringbuffer: /dev/shm/qb-cib_rw-response-25026-14152-13-header
Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
ringbuffer: /dev/shm/qb-cib_rw-event-25026-14152-13-header
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_setup.c:478 IPC
credentials authenticated (25004-14152-27)
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_shm.c:294 connecting
to client [14152]
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback:
Invoking the PE: query=39, ref=pe_calc-dc-1397290955-22, seq=1024, quorate=1
Apr 12 16:22:35 [25031] h1 crmd: debug:
qb_ipcs_dispatch_connection_request: HUP conn (25031-14152-14)
Apr 12 16:22:35 [25031] h1 crmd: debug: qb_ipcs_disconnect:
qb_ipcs_disconnect(25031-14152-14) state:2
Apr 12 16:22:35 [25031] h1 crmd: info: crm_client_destroy:
Destroying 0 events
Apr 12 16:22:35 [25031] h1 crmd: debug: qb_rb_close: Free'ing
ringbuffer: /dev/shm/qb-crmd-response-25031-14152-14-header
...
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH
timeout: 60000
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH
of failed nodes is disabled
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop
all active resources: false
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Cluster
is symmetric - resources can run anywhere by default
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Default
stickiness: 0
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On loss
of CCM Quorum: Stop ALL resources
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node
scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains:
Unpacking domains
Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
not allocate a node for VMdos-1
Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
VMdos-1 cannot run anywhere
Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
not allocate a node for VMdos-2
Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
VMdos-2 cannot run anywhere
Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-1
(Stopped)
Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-2
(Stopped)
Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message:
Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE
origin=handle_response ]
Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=handle_response ]
Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked
transition 5: 0 actions in 0 synapses
Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing
graph 5 (ref=pe_calc-dc-1397290955-22) derived from
/var/lib/pacemaker/pengine/pe-input-5.bz2
Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
transition graph
Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 5
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-5.bz2): Complete
Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
transition graph
Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger:
Transition 5 is now complete
Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing
transition completion in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 5
status: done - <null>
Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL
origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1 crmd: info: do_log: FSA: Input
I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition:
Starting PEngine Recheck Timer
Apr 12 16:22:35 [25031] h1 crmd: debug: crm_timer_start: Started
PEngine Recheck Timer (I_PE_CALC:900000ms), src=62
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents:
Archived previous version as /var/lib/pacemaker/cib/cib-9.raw
Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Writing
CIB to disk
Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
ringbuffer: /dev/shm/qb-cib_rw-request-25026-14152-13-header
Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents: Wrote
version 0.12.0 of the CIB to disk (digest: bd7d26226d6aa75f28b9eb670a67e944)
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:272 connection
created
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:306 lib_init_fn:
conn=0x7f32d0f4f4b0
Apr 12 16:22:35 [25024] h1 pacemakerd: notice: pcmk_ipc_dispatch:
Instructing peers to remove references to node (null)/1084752016
Apr 12 16:22:35 [25026] h1 cib: info: crm_client_new:
Connecting 0xa23770 for uid=0 gid=0 pid=14152
id=701dfa3c-d585-49d2-bd1e-f44636823e1b
Apr 12 16:22:35 [25026] h1 cib: debug: handle_new_connection:
IPC credentials authenticated (25026-14152-13)
Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_shm_connect:
connecting to client [14152]
Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Wrote
digest bd7d26226d6aa75f28b9eb670a67e944 to disk
Apr 12 16:22:35 [25026] h1 cib: info: retrieveCib: Reading cluster
configuration from: /var/lib/pacemaker/cib/cib.y6wnAg (digest:
/var/lib/pacemaker/cib/cib.ncXFvs)
...
Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents:
Activating /var/lib/pacemaker/cib/cib.y6wnAg
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_delete operation for section nodes: OK (rc=0,
origin=local/crm_node/2, version=0.12.2)
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:757 HUP conn
(25004-14152-27)
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:605
qb_ipcs_disconnect(25004-14152-27) state:2
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] loop_poll_epoll.c:117
epoll_ctl(del): Bad file descriptor (9)
Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:417
cs_ipcs_connection_closed()
Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:325 exit_fn for
conn=0x7f32d0f4f4b0
Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:390
cs_ipcs_connection_destroyed()
Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
Completed cib_delete operation for section status: OK (rc=0,
origin=local/crm_node/3, version=0.12.2)
...
Apr 12 16:22:35 [25024] h1 pacemakerd: debug:
qb_ipcs_dispatch_connection_request: HUP conn (25024-14152-10)
Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_disconnect:
qb_ipcs_disconnect(25024-14152-10) state:2
Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_destroy:
Destroying 0 events
...
Apr 12 16:22:35 [25026] h1 cib: debug:
qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect:
qb_ipcs_disconnect(25026-14152-13) state:2
Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy:
Destroying 0 events
...
Apr 12 16:22:35 [25024] h1 pacemakerd: notice: crm_reap_dead_member:
Removing h0/1084752016 from the membership list
Apr 12 16:22:35 [25024] h1 pacemakerd: notice: reap_crm_member: Purged
1 peers with id=1084752016 and/or uname=(null) from the membership cache
-------------------------------------------------------------------------------------------------------------------------------------
本邮件及其附件含有杭州华三通信技术有限公司的保密信息,仅限于发送给上面地址中列出
的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、
或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本
邮件!
This e-mail and its attachments contain confidential information from H3C,
which is
intended only for the person or entity whose address is listed above. Any use
of the
information contained herein in any way (including, but not limited to, total
or partial
disclosure, reproduction, or dissemination) by persons other than the intended
recipient(s) is prohibited. If you receive this e-mail in error, please notify
the sender
by phone or email immediately and delete it!
_______________________________________________
Pacemaker mailing list: [email protected]
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org