Open rotulet opened 5 years ago
Does nearly the same with versions:
pcs status gives:
Full list of resources:
Clone Set: DrbdVolume-clone [DrbdVolume]
Masters: [ carto-blade-0 ]
Started: [ carto-blade-1 ]
slave node is not detected as Slave in the 'Clone Set' and if I shutdown the master he does not promote the slave.
Here is the log if I stop the master:
crmd[30804]: notice: State transition S_IDLE -> S_POLICY_ENGINE
pengine[30803]: notice: On loss of CCM Quorum: Ignore
pengine[30803]: error: Resource start-up disabled since no STONITH resources have been defined
pengine[30803]: error: Either configure some or disable STONITH with the stonith-enabled option
pengine[30803]: error: NOTE: Clusters with shared data need STONITH to ensure data integrity
pengine[30803]: notice: Scheduling Node carto-blade-0 for shutdown
pengine[30803]: error: Couldn't expand DrbdVolume-clone_promote_0
pengine[30803]: notice: Calculated transition 4, saving inputs in /var/lib/pacemaker/pengine/pe-input-95.bz2
pengine[30803]: notice: Configuration ERRORs found during PE processing. Please run "crm_verify -L" to identify issues.
crmd[30804]: notice: Transition 4 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-95.bz2): Complete
crmd[30804]: notice: State transition S_TRANSITION_ENGINE -> S_IDLE
crmd[30804]: notice: do_shutdown of peer carto-blade-0 is complete
attrd[30802]: notice: Node carto-blade-0 state is now lost
attrd[30802]: notice: Removing all carto-blade-0 attributes for peer loss
attrd[30802]: notice: Purged 1 peers with id=1 and/or uname=carto-blade-0 from the membership cache
stonith-ng[30800]: notice: Node carto-blade-0 state is now lost
stonith-ng[30800]: notice: Purged 1 peers with id=1 and/or uname=carto-blade-0 from the membership cache
cib[30799]: notice: Node carto-blade-0 state is now lost
cib[30799]: notice: Purged 1 peers with id=1 and/or uname=carto-blade-0 from the membership cache
192.168.5.59: Stopping Cluster (pacemaker)...
corosync[30783]: [TOTEM ] A new membership (2:72) was formed. Members left: 1
corosync[30783]: [CPG ] downlist left_list: 1 received
corosync[30783]: [QUORUM] Members[1]: 2
corosync[30783]: [MAIN ] Completed service synchronization, ready to provide service.
crmd[30804]: notice: Node carto-blade-0 state is now lost
pacemakerd[30792]: notice: Node carto-blade-0 state is now lost
crmd[30804]: notice: do_shutdown of peer carto-blade-0 is complete
192.168.5.59: Stopping Cluster (corosync)...
corosync[30783]: [KNET ] link: host: 1 link: 0 is down
corosync[30783]: [KNET ] host: host: 1 has no active links
Affected Puppet, Ruby, OS and module versions/distributions
How to reproduce (e.g Puppet code you use)
What are you seeing
pcs status
show me my 2 nodes started (but no master nor slave):drbd seems happy:
ButI get theses errors in log:
What behaviour did you expect instead
If I do the same in bash (without puppet):
It works as intended:
Log output: