Hi! Obviously the logs around '2021-05-11 16:15:42 +02:00' would be interesting. Despite of that I'll have to try that drink named "quo rum" soon...;-)
Regards, Ulrich >>> <[email protected]> schrieb am 11.05.2021 um 16:43 in Nachricht <trinity-025ea1af-e35f-443f-812d-e45e0bd05b9f-1620744212201@3c-app-gmx-bs36>: > Hi, > > I'm using a CentOS 8.3.2011 with a pacemaker‑2.0.4‑6.el8_3.1.x86_64 + > corosync‑3.0.3‑4.el8.x86_64 and kmod‑drbd90‑9.0.25‑2.el8_3.elrepo.x86_64. > The cluster consists of two nodes which are providing a ha‑mariadb with the > help of two drbd devices for the database and the logfiles. The corosync is > working over two rings and both machines are virtual kvm‑guests. > > Problem: > Node susanne is the active node and lisbon is changing from standby to > active, susanna is trying to demote one drbd‑device but is failling to. The > cluster is working on properly, but the error stays. > This is the what happens: > > Cluster Summary: > * Stack: corosync > * Current DC: lisbon (version 2.0.4‑6.el8_3.1‑2deceaa3ae) ‑ partition with > quo rum > * Last updated: Tue May 11 16:15:54 2021 > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > * 2 nodes configured > * 11 resource instances configured > > Node List: > * Online: [ lisbon susanne ] > > Active Resources: > * HA_IP (ocf::heartbeat:IPaddr2): Started susanne > * Clone Set: database_drbd‑clone [database_drbd] (promotable): > * Masters: [ susanne ] > * Slaves: [ lisbon ] > * Clone Set: drbd_logsfiles‑clone [drbd_logsfiles] (promotable): > * drbd_logsfiles (ocf::linbit:drbd): Demoting susanne > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > * fs_database (ocf::heartbeat:Filesystem): Started susanne > * mysql‑server (ocf::heartbeat:mysql): Started susanne > * Clone Set: ping_fw‑clone [ping_fw]: > * Started: [ lisbon susanne ] > > ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑ > after a few seconds it switches over: > > Cluster Summary: > * Stack: corosync > * Current DC: lisbon (version 2.0.4‑6.el8_3.1‑2deceaa3ae) ‑ partition with > quo rum > * Last updated: Tue May 11 16:17:59 2021 > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > * 2 nodes configured > * 11 resource instances configured > > Node List: > * Online: [ lisbon susanne ] > > Active Resources: > * HA_IP (ocf::heartbeat:IPaddr2): Started susanne > * Clone Set: database_drbd‑clone [database_drbd] (promotable): > * Masters: [ susanne ] > * Slaves: [ lisbon ] > * Clone Set: drbd_logsfiles‑clone [drbd_logsfiles] (promotable): > * Masters: [ susanne ] > * Slaves: [ lisbon ] > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > * fs_database (ocf::heartbeat:Filesystem): Started susanne > * mysql‑server (ocf::heartbeat:mysql): Started susanne > * Resource Group: apache: > * httpd_srv (ocf::heartbeat:apache): Started susanne > * Clone Set: ping_fw‑clone [ping_fw]: > * Started: [ lisbon susanne ] > > Failed Resource Actions: > * drbd_logsfiles_demote_0 on susanne 'error' (1): call=736, status='Timed > Out' > , exitreason='', last‑rc‑change='2021‑05‑11 16:15:42 +02:00', queued=0ms, exec=9 > 0001ms > ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑ > > I think it is a constraint‑problem, but I can't find it. > This is my config: > [root@susanne pacemaker]# pcs config show Cluster Name: mysql_cluster > Corosync Nodes: > susanne lisbon > Pacemaker Nodes: > lisbon susanne > > Resources: > Resource: HA_IP (class=ocf provider=heartbeat type=IPaddr2) > Attributes: cidr_netmask=24 ip=192.168.18.154 > Operations: monitor interval=15s (HA_IP‑monitor‑interval‑15s) > start interval=0s timeout=20s (HA_IP‑start‑interval‑0s) > stop interval=0s timeout=20s (HA_IP‑stop‑interval‑0s) > Clone: database_drbd‑clone > Meta Attrs: clone‑max=2 clone‑node‑max=1 notify=true promotable=true > promoted‑max=1 promoted‑node‑max=1 > Resource: database_drbd (class=ocf provider=linbit type=drbd) > Attributes: drbd_resource=drbd1 > Operations: demote interval=0s timeout=90 (database_drbd‑demote‑interval‑0s) > monitor interval=20 role=Slave timeout=20 > (database_drbd‑monitor‑interval‑20) > monitor interval=10 role=Master timeout=20 > (database_drbd‑monitor‑interval‑10) > notify interval=0s timeout=90 (database_drbd‑notify‑interval‑0s) > promote interval=0s timeout=90 > (database_drbd‑promote‑interval‑0s) > reload interval=0s timeout=30 (database_drbd‑reload‑interval‑0s) > start interval=0s timeout=240 (database_drbd‑start‑interval‑0s) > stop interval=0s timeout=100 (database_drbd‑stop‑interval‑0s) > Clone: drbd_logsfiles‑clone > Meta Attrs: clone‑max=2 clone‑node‑max=1 notify=true promotable=true > promoted‑max=1 promoted‑node‑max=1 > Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd) > Attributes: drbd_resource=drbd2 > Operations: demote interval=0s timeout=90 > (drbd_logsfiles‑demote‑interval‑0s) > monitor interval=20 role=Slave timeout=20 > (drbd_logsfiles‑monitor‑interval‑20) > monitor interval=10 role=Master timeout=20 > (drbd_logsfiles‑monitor‑interval‑10) > notify interval=0s timeout=90 > (drbd_logsfiles‑notify‑interval‑0s) > promote interval=0s timeout=90 > (drbd_logsfiles‑promote‑interval‑0s) > reload interval=0s timeout=30 > (drbd_logsfiles‑reload‑interval‑0s) > start interval=0s timeout=240 (drbd_logsfiles‑start‑interval‑0s) > stop interval=0s timeout=100 (drbd_logsfiles‑stop‑interval‑0s) > Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem) > Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4 > Operations: monitor interval=20s timeout=40s > (fs_logfiles‑monitor‑interval‑20s) > start interval=0s timeout=60s (fs_logfiles‑start‑interval‑0s) > stop interval=0s timeout=60s (fs_logfiles‑stop‑interval‑0s) > Resource: fs_database (class=ocf provider=heartbeat type=Filesystem) > Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4 > Operations: monitor interval=20s timeout=40s > (fs_database‑monitor‑interval‑20s) > start interval=0s timeout=60s (fs_database‑start‑interval‑0s) > stop interval=0s timeout=60s (fs_database‑stop‑interval‑0s) > Resource: mysql‑server (class=ocf provider=heartbeat type=mysql) > Attributes: additional_parameters=‑‑bind‑address=0.0.0.0 > binary=/usr/bin/mysqld_safe config=/etc/my.cnf datadir=/mnt/clusterfs1/mysql > pid=/var/lib/mysql/run/mariadb.pid socket=/var/lib/mysql/mysql.sock > Operations: demote interval=0s timeout=120s (mysql‑server‑demote‑interval‑0s) > monitor interval=20s timeout=30s > (mysql‑server‑monitor‑interval‑20s) > notify interval=0s timeout=90s (mysql‑server‑notify‑interval‑0s) > promote interval=0s timeout=120s > (mysql‑server‑promote‑interval‑0s) > start interval=0s timeout=60s (mysql‑server‑start‑interval‑0s) > stop interval=0s timeout=60s (mysql‑server‑stop‑interval‑0s) > Group: apache > Resource: httpd_srv (class=ocf provider=heartbeat type=apache) > Attributes: configfile=/etc/httpd/conf/httpd.conf > statusurl=http://127.0.0.1/server‑status > Operations: monitor interval=10s timeout=20s > (httpd_srv‑monitor‑interval‑10s) > start interval=0s timeout=40s (httpd_srv‑start‑interval‑0s) > stop interval=0s timeout=60s (httpd_srv‑stop‑interval‑0s) > Clone: ping_fw‑clone > Resource: ping_fw (class=ocf provider=pacemaker type=ping) > Attributes: dampen=10s host_list=192.168.18.1 multiplier=1000 > Operations: monitor interval=10s timeout=60s (ping_fw‑monitor‑interval‑10s) > start interval=0s timeout=60s (ping_fw‑start‑interval‑0s) > stop interval=0s timeout=20s (ping_fw‑stop‑interval‑0s) > > Stonith Devices: > Fencing Levels: > > Location Constraints: > Resource: mysql‑server > Constraint: location‑mysql‑server > Rule: boolean‑op=or score=‑INFINITY (id:location‑mysql‑server‑rule) > Expression: pingd lt 1 (id:location‑mysql‑server‑rule‑expr) > Expression: not_defined pingd (id:location‑mysql‑server‑rule‑expr‑1) > Ordering Constraints: > start mysql‑server then start httpd_srv (kind:Mandatory) > (id:order‑mysql‑server‑httpd_srv‑mandatory) > start database_drbd‑clone then start drbd_logsfiles‑clone (kind:Mandatory) > (id:order‑database_drbd‑clone‑drbd_logsfiles‑clone‑mandatory) > start drbd_logsfiles‑clone then start fs_database (kind:Mandatory) > (id:order‑drbd_logsfiles‑clone‑fs_database‑mandatory) > start fs_database then start fs_logfiles (kind:Mandatory) > (id:order‑fs_database‑fs_logfiles‑mandatory) > start fs_logfiles then start mysql‑server (kind:Mandatory) > (id:order‑fs_logfiles‑mysql‑server‑mandatory) > Colocation Constraints: > fs_logfiles with drbd_logsfiles‑clone (score:INFINITY) (with‑rsc‑role:Master) > (id:colocation‑fs_logfiles‑drbd_logsfiles‑clone‑INFINITY) > fs_database with database_drbd‑clone (score:INFINITY) (with‑rsc‑role:Master) > (id:colocation‑fs_database‑database_drbd‑clone‑INFINITY) > drbd_logsfiles‑clone with database_drbd‑clone (score:INFINITY) > (rsc‑role:Master) (with‑rsc‑role:Master) > (id:colocation‑drbd_logsfiles‑clone‑database_drbd‑clone‑INFINITY) > HA_IP with database_drbd‑clone (score:INFINITY) (rsc‑role:Started) > (with‑rsc‑role:Master) (id:colocation‑HA_IP‑database_drbd‑clone‑INFINITY) > mysql‑server with fs_database (score:INFINITY) > (id:colocation‑mysql‑server‑fs_database‑INFINITY) > httpd_srv with mysql‑server (score:INFINITY) > (id:colocation‑httpd_srv‑mysql‑server‑INFINITY) > Ticket Constraints: > > Alerts: > No alerts defined > > Resources Defaults: > No defaults set > Operations Defaults: > No defaults set > > Cluster Properties: > cluster‑infrastructure: corosync > cluster‑name: mysql_cluster > dc‑version: 2.0.4‑6.el8_3.1‑2deceaa3ae > have‑watchdog: false > last‑lrm‑refresh: 1620742514 > stonith‑enabled: FALSE > > Tags: > No tags defined > > Quorum: > Options: > > > > > > Any suggestions are welcome > > best regards stay safe, take care > > fatcharly > > _______________________________________________ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ _______________________________________________ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
