On 13 Apr 2014, at 11:36 am, Caizhifeng <[email protected]> wrote: > Hi all, > I'm kinda new to Pacemaker and have some questions I hope some of you could > help me with, any idea(s) would be highly appreciated, thank you. > I'am building the HA cluster with corosync2.3.3 + cluster-glue + > pacemaker-1.1.11-rc5. The problem is, when I try to remove a node from HA > cluster with "crm_node -f -R nodeID", it seems to remove the wrong node. > > My setup is as follows(and it can be reproduced): > 1. The HA is ok with 2 resource and 2 nodes, and with stonith-enabled="false": > root@h1:/opt/bin# crm status > Last updated: Sat Apr 12 16:00:51 2014 > Last change: Sat Apr 12 16:00:43 2014 via cibadmin on h1 > Stack: corosync > Current DC: h1 (1084752017) - partition with quorum > Version: 1.1.10-33f9d09 > 2 Nodes configured > 2 Resources configured > > > Online: [ h0 h1 ] > > VMdos-1 (ocf::heartbeat:VirtualDomain): Started h1 > root@h1:/opt/bin# > > root@h1:/opt/bin# crm_node --list > 1084752017 h1 > 1084752016 h0 > root@h1:/opt/bin# > > 2. remove node h0 from HA > (1). Stop pacemakerd and corosync service on node h0 (2). Run the follow cmd > on node h1 to Remove node info from cib: > cibadmin --delete --obj_type status --crm_xml "<node_state > id=\"1084752016\"/>" > cibadmin --delete --obj_type nodes --crm_xml "<node id=\"1084752016\"/>" > > at this time; the result of "crm status" shows that node " h0 " is offline > and there's no information of node h0 in CIB. But the "crm_node --list " > still include node" h0" > > 3. crm_node -f -R 1084752016
Could you re-run this command but add -VVVVVVVV please?
I suspect the fault lies in this binary.
> After this cmd, node " h1" is removed, and result of "crm status " is as
> follow(with 0 node):
>
> root@h1:/opt/bin# crm status
> Last updated: Sat Apr 12 15:59:42 2014
> Last change: Sat Apr 12 15:59:37 2014 via crm_node on h1
> Stack: corosync
> Current DC: NONE
> 0 Nodes configured
> 2 Resources configured
> root@h1:/opt/bin
>
>
> The corosync.conf is as follow:
> .............
> quorum {
> provider: corosync_votequorum
> expected_votes: 2
> allow_downscale: 1
> two_node: 1
> }
> ...........
>
>
> The corosync.log is as follow, it seems node " h1" is removed due to fencing!
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:299
> Free'ing ringbuffer: /dev/shm/qb-cmap-request-25004-14152-27-header
> Apr 12 16:22:35 [25026] h1 cib: debug: activateCibXml:
> Triggering CIB write for cib_delete op
> Apr 12 16:22:35 [25031] h1 crmd: notice: crm_reap_dead_member:
> Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25031] h1 crmd: notice: reap_crm_member:
> Purged 1 peers with id=1084752016 and/or uname=(null) from the membership
> cache
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: log_cib_diff: Config
> update: Local-only Change: 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> <configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> <nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> </nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> </configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++
> <cib epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
> update-origin="h1" update-client="crm_node" have-quorum="1"
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1 cib: notice: log_cib_diff: cib:diff:
> Local-only Change: 0.12.1
> Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: --
> <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: ++ <cib
> epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
> update-origin="h1" update-client="crm_node" have-quorum="1"
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_delete operation for section nodes: OK (rc=0,
> origin=local/crm_node/2, version=0.12.1)
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff:
> Processing diff (cib_delete): 0.11.1 -> 0.12.1 (S_IDLE)
> Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph:
> te_update_diff:126 - Triggered transition abort (complete=1, node=,
> tag=diff, id=(null), magic=NA, cib=0.12.1) : Non-status change
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <diff crm_feature_set="3.0.8"
> digest="3cccef06483ac4dfeadfb562f6f8293a">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <diff-removed admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <configuration>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <nodes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <node id="1084752017" uname="h1"
> __crm_diff_marker__="removed:top"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </nodes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </configuration>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </cib>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </diff-removed>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <diff-added>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <cib epoch="12" num_updates="1" admin_epoch="0"
> validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat
> Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node"
> have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </diff-added>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </diff>
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
> I_PE_CALC: [ state=S_IDLE cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State
> transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
> origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1
> cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 38:
> Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_query operation for section 'all': OK (rc=0,
> origin=local/crmd/38, version=0.12.1)
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback:
> Invoking the PE: query=38, ref=pe_calc-dc-1397290955-21, seq=1024, quorate=1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff:
> --- 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff:
> +++ 0.12.2 3d673c27c3c92939b41c7207edee9f46
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> <cib num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> <status>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <node_state id="1084752017" uname="h1" in_ccm="true" crmd="online"
> crm-debug-origin="post_cache_update" join="member" expected="member">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <nvpair id="status-1084752017-probe_complete" name="probe_complete"
> value="true"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </instance_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </transient_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_resource id="VMdos-1" type="VirtualDomain" class="ocf"
> provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_rsc_op id="VMdos-1_last_failure_0"
> operation_key="VMdos-1_monitor_0" operation="monitor"
> crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8"
> transition-key="6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c"
> transition-magic="0:0;6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="5"
> rc-code="0" op-status="0" interval="0" last-run="1397290607"
> last-rc-change="1397290607" exec-time="92" queue-time="0" op-digest="
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_rsc_op id="VMdos-1_monitor_30000"
> operation_key="VMdos-1_monitor_30000" operation="monitor"
> crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8"
> transition-key="7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c"
> transition-magic="0:0;7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="6"
> rc-code="0" op-status="0" interval="30000" last-rc-change="1397290607"
> exec-time="67" queue-time="0" op-digest="0874c7ce5f61a12
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_resource id="VMdos-2" type="VirtualDomain" class="ocf"
> provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> <lrm_rsc_op id="VMdos-2_last_0" operation_key="VMdos-2_monitor_0"
> operation="monitor" crm-debug-origin="build_active_RAs"
> crm_feature_set="3.0.8"
> transition-key="7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c"
> transition-magic="0:7;7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c"
> call-id="10" rc-code="7" op-status="0" interval="0" last-run="1397290608"
> last-rc-change="1397290608" exec-time="61" queue-time="0" op-digest="c7d22be
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </lrm>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: --
> </node_state>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> </status>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -
> </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++
> <cib epoch="12" num_updates="2" admin_epoch="0" validate-with="pacemaker-1.2"
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014"
> update-origin="h1" update-client="crm_node" have-quorum="1"
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop
> all active resources: false
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> Default stickiness: 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On
> loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node
> scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains:
> Unpacking domains
> Apr 12 16:22:35 [25030] h1 pengine: warning: unpack_status: Node
> h1 in status section no longer exists
> Apr 12 16:22:35 [25030] h1 pengine: info: unpack_status: Node
> 1084752017 is unknown
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
> not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
> VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
> not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
> VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave
> VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave
> VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message:
> Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
> I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE
> origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
> transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
> cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked
> transition 4: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing
> graph 4 (ref=pe_calc-dc-1397290955-21) derived from
> /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
> transition graph
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_delete operation for section status: OK (rc=0,
> origin=local/crm_node/3, version=0.12.2)
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff:
> Processing diff (cib_delete): 0.12.1 -> 0.12.2 (S_TRANSITION_ENGINE)
> Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph:
> te_update_diff:188 - Triggered transition abort (complete=0, node=h1,
> tag=transient_attributes, id=1084752017, magic=NA, cib=0.12.2) : Transient
> attribute: removal
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <nvpair id="status-1084752017-shutdown" name="shutdown"
> value="0"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause <nvpair id="status-1084752017-probe_complete"
> name="probe_complete" value="true"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </instance_attributes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph:
> Cause </transient_attributes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority:
> Abort priority upgraded from 0 to 1000000
> Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority:
> Abort action done superceeded by restart
> Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 4
> (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
> transition graph
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger:
> Transition 4 is now complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing
> transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 4
> status: restart - Transient attribute: removal
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
> I_PE_CALC: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL
> origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
> transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC
> cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1
> cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 39:
> Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_new:
> Connecting 0x25b4ea0 for uid=0 gid=0 pid=14152
> id=f3612e17-0806-4355-a3fc-2cf1feda1e6d
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: handle_new_connection:
> IPC credentials authenticated (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_shm_connect:
> connecting to client [14152]
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_query operation for section 'all': OK (rc=0,
> origin=local/crmd/39, version=0.12.2)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25026] h1 cib: debug:
> qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect:
> qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy:
> Destroying 0 events
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
> ringbuffer: /dev/shm/qb-cib_rw-response-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
> ringbuffer: /dev/shm/qb-cib_rw-event-25026-14152-13-header
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_setup.c:478 IPC
> credentials authenticated (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_shm.c:294 connecting
> to client [14152]
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback:
> Invoking the PE: query=39, ref=pe_calc-dc-1397290955-22, seq=1024, quorate=1
> Apr 12 16:22:35 [25031] h1 crmd: debug:
> qb_ipcs_dispatch_connection_request: HUP conn (25031-14152-14)
> Apr 12 16:22:35 [25031] h1 crmd: debug: qb_ipcs_disconnect:
> qb_ipcs_disconnect(25031-14152-14) state:2
> Apr 12 16:22:35 [25031] h1 crmd: info: crm_client_destroy:
> Destroying 0 events
> Apr 12 16:22:35 [25031] h1 crmd: debug: qb_rb_close: Free'ing
> ringbuffer: /dev/shm/qb-crmd-response-25031-14152-14-header
> ...
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop
> all active resources: false
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config:
> Default stickiness: 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On
> loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node
> scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains:
> Unpacking domains
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
> not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
> VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could
> not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource
> VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave
> VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave
> VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message:
> Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
> I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE
> origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State
> transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
> cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked
> transition 5: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing
> graph 5 (ref=pe_calc-dc-1397290955-22) derived from
> /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
> transition graph
> Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 5
> (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> Source=/var/lib/pacemaker/pengine/pe-input-5.bz2): Complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty
> transition graph
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger:
> Transition 5 is now complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing
> transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 5
> status: done - <null>
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing
> I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL
> origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_log: FSA: Input
> I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State
> transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition:
> Starting PEngine Recheck Timer
> Apr 12 16:22:35 [25031] h1 crmd: debug: crm_timer_start:
> Started PEngine Recheck Timer (I_PE_CALC:900000ms), src=62
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents:
> Archived previous version as /var/lib/pacemaker/cib/cib-9.raw
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents:
> Writing CIB to disk
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing
> ringbuffer: /dev/shm/qb-cib_rw-request-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents: Wrote
> version 0.12.0 of the CIB to disk (digest: bd7d26226d6aa75f28b9eb670a67e944)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:272
> connection created
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:306 lib_init_fn:
> conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: pcmk_ipc_dispatch:
> Instructing peers to remove references to node (null)/1084752016
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_new:
> Connecting 0xa23770 for uid=0 gid=0 pid=14152
> id=701dfa3c-d585-49d2-bd1e-f44636823e1b
> Apr 12 16:22:35 [25026] h1 cib: debug: handle_new_connection:
> IPC credentials authenticated (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_shm_connect:
> connecting to client [14152]
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Wrote
> digest bd7d26226d6aa75f28b9eb670a67e944 to disk
> Apr 12 16:22:35 [25026] h1 cib: info: retrieveCib: Reading
> cluster configuration from: /var/lib/pacemaker/cib/cib.y6wnAg (digest:
> /var/lib/pacemaker/cib/cib.ncXFvs)
> ...
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents:
> Activating /var/lib/pacemaker/cib/cib.y6wnAg
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_delete operation for section nodes: OK (rc=0,
> origin=local/crm_node/2, version=0.12.2)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:757 HUP conn
> (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:605
> qb_ipcs_disconnect(25004-14152-27) state:2
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] loop_poll_epoll.c:117
> epoll_ctl(del): Bad file descriptor (9)
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:417
> cs_ipcs_connection_closed()
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:325 exit_fn for
> conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:390
> cs_ipcs_connection_destroyed()
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request:
> Completed cib_delete operation for section status: OK (rc=0,
> origin=local/crm_node/3, version=0.12.2)
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug:
> qb_ipcs_dispatch_connection_request: HUP conn (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_disconnect:
> qb_ipcs_disconnect(25024-14152-10) state:2
> Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_destroy:
> Destroying 0 events
> ...
> Apr 12 16:22:35 [25026] h1 cib: debug:
> qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect:
> qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy:
> Destroying 0 events
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: crm_reap_dead_member:
> Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: reap_crm_member:
> Purged 1 peers with id=1084752016 and/or uname=(null) from the membership
> cache
>
>
> -------------------------------------------------------------------------------------------------------------------------------------
> 本邮件及其附件含有杭州华三通信技术有限公司的保密信息,仅限于发送给上面地址中列出
> 的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、
> 或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本
> 邮件!
> This e-mail and its attachments contain confidential information from H3C,
> which is
> intended only for the person or entity whose address is listed above. Any use
> of the
> information contained herein in any way (including, but not limited to, total
> or partial
> disclosure, reproduction, or dissemination) by persons other than the intended
> recipient(s) is prohibited. If you receive this e-mail in error, please
> notify the sender
> by phone or email immediately and delete it!
> _______________________________________________
> Pacemaker mailing list: [email protected]
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ Pacemaker mailing list: [email protected] http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
