From mboxrd@z Thu Jan  1 00:00:00 1970
Return-Path: <aderumier@odiso.com>
Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68])
 (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits)
 key-exchange X25519 server-signature RSA-PSS (2048 bits))
 (No client certificate requested)
 by lists.proxmox.com (Postfix) with ESMTPS id 9F57E620BA
 for <pve-devel@lists.proxmox.com>; Tue, 15 Sep 2020 17:59:10 +0200 (CEST)
Received: from firstgate.proxmox.com (localhost [127.0.0.1])
 by firstgate.proxmox.com (Proxmox) with ESMTP id 82D221C3A3
 for <pve-devel@lists.proxmox.com>; Tue, 15 Sep 2020 17:58:40 +0200 (CEST)
Received: from mailpro.odiso.net (mailpro.odiso.net [89.248.211.110])
 (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits))
 (No client certificate requested)
 by firstgate.proxmox.com (Proxmox) with ESMTPS id 396B61C382
 for <pve-devel@lists.proxmox.com>; Tue, 15 Sep 2020 17:58:34 +0200 (CEST)
Received: from localhost (localhost [127.0.0.1])
 by mailpro.odiso.net (Postfix) with ESMTP id E74C01A5FCB9;
 Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
Received: from mailpro.odiso.net ([127.0.0.1])
 by localhost (mailpro.odiso.net [127.0.0.1]) (amavisd-new, port 10032)
 with ESMTP id OalT2H2EV0CI; Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
Received: from localhost (localhost [127.0.0.1])
 by mailpro.odiso.net (Postfix) with ESMTP id CB1861A5FCC6;
 Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
X-Virus-Scanned: amavisd-new at mailpro.odiso.com
Received: from mailpro.odiso.net ([127.0.0.1])
 by localhost (mailpro.odiso.net [127.0.0.1]) (amavisd-new, port 10026)
 with ESMTP id RqbKcW7ZyXqg; Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
Received: from mailpro.odiso.net (mailpro.odiso.net [10.1.31.111])
 by mailpro.odiso.net (Postfix) with ESMTP id A56351A5FCC5;
 Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
Date: Tue, 15 Sep 2020 17:58:33 +0200 (CEST)
From: Alexandre DERUMIER <aderumier@odiso.com>
To: Proxmox VE development discussion <pve-devel@lists.proxmox.com>
Cc: Thomas Lamprecht <t.lamprecht@proxmox.com>
Message-ID: <597522514.840749.1600185513450.JavaMail.zimbra@odiso.com>
In-Reply-To: <132388307.839866.1600181866529.JavaMail.zimbra@odiso.com>
References: <216436814.339545.1599142316781.JavaMail.zimbra@odiso.com>
 <1282130277.831843.1600164947209.JavaMail.zimbra@odiso.com>
 <1732268946.834480.1600167871823.JavaMail.zimbra@odiso.com>
 <1800811328.836757.1600174194769.JavaMail.zimbra@odiso.com>
 <43250fdc-55ba-03d9-2507-a2b08c5945ce@proxmox.com>
 <1798333820.838842.1600178990068.JavaMail.zimbra@odiso.com>
 <6b680921-12d0-006b-6d04-bbe1c4bb04f8@proxmox.com>
 <132388307.839866.1600181866529.JavaMail.zimbra@odiso.com>
MIME-Version: 1.0
Content-Type: text/plain; charset=utf-8
Content-Transfer-Encoding: quoted-printable
X-Mailer: Zimbra 8.8.12_GA_3866 (ZimbraWebClient - GC83 (Linux)/8.8.12_GA_3844)
Thread-Topic: corosync bug: cluster break after 1 node clean shutdown
Thread-Index: fqzQ8CV4gT3UroNXiJlm8US/HHWe/CRQ8Ou9
X-SPAM-LEVEL: Spam detection results:  0
 AWL 0.088 Adjusted score from AWL reputation of From: address
 KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment
 RCVD_IN_DNSWL_NONE     -0.0001 Sender listed at https://www.dnswl.org/,
 no trust
 SPF_HELO_NONE           0.001 SPF: HELO does not publish an SPF Record
 SPF_PASS               -0.001 SPF: sender matches SPF record
Subject: Re: [pve-devel] corosync bug: cluster break after 1 node clean
 shutdown
X-BeenThere: pve-devel@lists.proxmox.com
X-Mailman-Version: 2.1.29
Precedence: list
List-Id: Proxmox VE development discussion <pve-devel.lists.proxmox.com>
List-Unsubscribe: <https://lists.proxmox.com/cgi-bin/mailman/options/pve-devel>, 
 <mailto:pve-devel-request@lists.proxmox.com?subject=unsubscribe>
List-Archive: <http://lists.proxmox.com/pipermail/pve-devel/>
List-Post: <mailto:pve-devel@lists.proxmox.com>
List-Help: <mailto:pve-devel-request@lists.proxmox.com?subject=help>
List-Subscribe: <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel>, 
 <mailto:pve-devel-request@lists.proxmox.com?subject=subscribe>
X-List-Received-Date: Tue, 15 Sep 2020 15:59:10 -0000

Another small lock at 17:41:09

To be sure, I have done a small loop of write each second in /etc/pve,  nod=
e node2.

it's hanging at first corosync restart,   then, on second corosync restart =
it's working again.

I'll try to improve this tomorrow to be able to debug corosync process
- restarting corosync
  do some write in /etc/pve/
 - and if it's hanging don't restart corosync again



node2: echo test > /etc/pve/test loop
--------------------------------------
Current time : 17:41:01
Current time : 17:41:02
Current time : 17:41:03
Current time : 17:41:04
Current time : 17:41:05
Current time : 17:41:06
Current time : 17:41:07
Current time : 17:41:08
Current time : 17:41:09

hang

Current time : 17:42:05
Current time : 17:42:06
Current time : 17:42:07



node1
-----
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] pmtud: PMTUD completed f=
or host: 6 link: 0 current link mtu: 1397
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] pmtud: Starting PMTUD fo=
r host: 10 link: 0
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] udp: detected kernel MTU=
: 1500
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [TOTEM ] Knet pMTU change: 1397
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] pmtud: PMTUD link change=
 for host: 10 link: 0 from 469 to 1397
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] pmtud: PMTUD completed f=
or host: 10 link: 0 current link mtu: 1397
Sep 15 17:41:08 m6kvm1 corosync[18145]:   [KNET  ] pmtud: Global data MTU c=
hanged to: 1397
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] IPC credentials authenti=
cated (/dev/shm/qb-18145-16239-31-zx6KJM/qb)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] connecting to client [16=
239]
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] connection created
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QUORUM] lib_init_fn: conn=3D0x55=
6c2918d5f0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QUORUM] got quorum_type request =
on 0x556c2918d5f0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QUORUM] got trackstart request o=
n 0x556c2918d5f0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QUORUM] sending initial status t=
o 0x556c2918d5f0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QUORUM] sending quorum notificat=
ion to 0x556c2918d5f0, length =3D 52
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] IPC credentials authenti=
cated (/dev/shm/qb-18145-16239-32-I7ZZ6e/qb)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] connecting to client [16=
239]
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] connection created
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CMAP  ] lib_init_fn: conn=3D0x55=
6c2918ef20
Sep 15 17:41:09 m6kvm1 pmxcfs[16239]: [status] notice: update cluster info =
(cluster name  m6kvm, version =3D 20)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] IPC credentials authenti=
cated (/dev/shm/qb-18145-16239-33-6RKbvH/qb)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] connecting to client [16=
239]
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] connection created
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] lib_init_fn: conn=3D0x55=
6c2918ad00, cpd=3D0x556c2918b50c
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] IPC credentials authenti=
cated (/dev/shm/qb-18145-16239-34-GAY5T9/qb)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] connecting to client [16=
239]
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [QB    ] shm size:1048589; real_s=
ize:1052672; rb->word_size:263168
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] connection created
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] lib_init_fn: conn=3D0x55=
6c2918c740, cpd=3D0x556c2918ce8c
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] Creating commit token be=
cause I am the rep.
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] Saving state aru 5 high =
seq received 5
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Storing new sequence id =
for ring 1197
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] entering COMMIT state.
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] got commit token
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] entering RECOVERY state.
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] TRANS [0] member 1:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [0] member 1:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (1.1193)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 5 high delivered 5 r=
eceived flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [1] member 2:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [2] member 3:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [3] member 4:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
ep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 12=
3 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [4] member 5:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [5] member 6:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [6] member 7:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [7] member 8:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [8] member 9:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [9] member 10:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [10] member 11:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [11] member 12:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [12] member 13:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] position [13] member 14:
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] Did not need to originat=
e any messages in recovery.
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] got commit token
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] Sending initial ORF toke=
n
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 0, aru 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 1, aru 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 2, aru 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 3, aru 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] retrans flag count 4 tok=
en aru 0 install seq 0 aru 0 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] Resetting old ring state
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] recovery to regular 1-0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] waiting_trans_ack change=
d to 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.90)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.91)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.92)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.93)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.94)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.95)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.96)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.97)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.107)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.108)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.109)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.110)=20
ep 15 17:41:09 m6kvm1 corosync[18145]:   [MAIN  ] Member joined: r(0) ip(10=
.3.94.111)=20
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [SYNC  ] call init for locally kn=
own services
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] entering OPERATIONAL sta=
te.
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [TOTEM ] A new membership (1.1197=
) was formed. Members joined: 2 3 4 5 6 7 8 9 10 11 12 13 14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [SYNC  ] enter sync process
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [SYNC  ] Committing synchronizati=
on for corosync configuration map access
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 2
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 3
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 4
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 5
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 6
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 7
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 8
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 9
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 10
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 11
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 12
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] got joinlist message fro=
m node 13
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [SYNC  ] Committing synchronizati=
on for corosync cluster closed process group service v1.01
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] my downlist: members(old=
:1 left:0)
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[0] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.110) , pid:30209
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[1] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.110) , pid:30209
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[2] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.109) , pid:31350
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[3] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.109) , pid:31350
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[4] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.108) , pid:3569
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[5] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.108) , pid:3569
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[6] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.107) , pid:19504
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[7] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.107) , pid:19504
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[8] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.97) , pid:11947
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[9] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.97) , pid:11947
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[10] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.96) , pid:20814
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[11] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.96) , pid:20814
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[12] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.95) , pid:39420
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[13] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.95) , pid:39420
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[14] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.94) , pid:12452
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[15] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.94) , pid:12452
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[16] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.93) , pid:44300
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[17] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.93) , pid:44300
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[18] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.92) , pid:42259
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[19] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.92) , pid:42259
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[20] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.91) , pid:40630
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[21] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.91) , pid:40630
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[22] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.90) , pid:25870
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[23] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.90) , pid:25870
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[24] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.111) , pid:25634
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [CPG   ] joinlist_messages[25] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.111) , pid:25634
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] flags: quorate: No Leavi=
ng: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVot=
e: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] Sending nodelist callbac=
k. ring_id =3D 1.1197
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 13
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[13]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] total_votes=3D2, expecte=
d_votes=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 13
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[14]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] total_votes=3D3, expecte=
d_votes=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[1]: vot=
es: 1, expected: 14 flags: 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] flags: quorate: No Leavi=
ng: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVot=
e: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] total_votes=3D3, expecte=
d_votes=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 2
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[2]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] total_votes=3D4, expecte=
d_votes=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 2 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 2
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm1 corosync[18145]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 3
....
....
next corosync restart

Sep 15 17:42:03 m6kvm1 corosync[18145]:   [MAIN  ] Node was shut down by a =
signal
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [SERV  ] Unloading all Corosync s=
ervice engines.
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [QB    ] withdrawing server socke=
ts
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [QB    ] qb_ipcs_unref() - destro=
ying
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [SERV  ] Service engine unloaded:=
 corosync vote quorum service v1.0
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [QB    ] qb_ipcs_disconnect(/dev/=
shm/qb-18145-16239-32-I7ZZ6e/qb) state:2
Sep 15 17:42:03 m6kvm1 pmxcfs[16239]: [confdb] crit: cmap_dispatch failed: =
2
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [MAIN  ] cs_ipcs_connection_close=
d()=20
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [CMAP  ] exit_fn for conn=3D0x556=
c2918ef20
Sep 15 17:42:03 m6kvm1 corosync[18145]:   [MAIN  ] cs_ipcs_connection_destr=
oyed()=20


node2
-----



Sep 15 17:41:05 m6kvm2 corosync[25411]:   [KNET  ] pmtud: Starting PMTUD fo=
r host: 10 link: 0
Sep 15 17:41:05 m6kvm2 corosync[25411]:   [KNET  ] udp: detected kernel MTU=
: 1500
Sep 15 17:41:05 m6kvm2 corosync[25411]:   [KNET  ] pmtud: PMTUD completed f=
or host: 10 link: 0 current link mtu: 1397
Sep 15 17:41:07 m6kvm2 corosync[25411]:   [KNET  ] rx: host: 1 link: 0 rece=
ived pong: 2
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [TOTEM ] entering GATHER state fr=
om 11(merge during join).
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:08 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: Source host 1 not re=
achable yet. Discarding packet.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] rx: host: 1 link: 0 is u=
p
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] Knet host change callbac=
k. nodeid: 1 reachable: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] host: host: 1 (passive) =
best link: 0 (pri: 1)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] pmtud: Starting PMTUD fo=
r host: 1 link: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] udp: detected kernel MTU=
: 1500
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [KNET  ] pmtud: PMTUD completed f=
or host: 1 link: 0 current link mtu: 1397
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] got commit token
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] Saving state aru 123 hig=
h seq received 123
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [MAIN  ] Storing new sequence id =
for ring 1197
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] entering COMMIT state.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] got commit token
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] entering RECOVERY state.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [0] member 2:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [1] member 3:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [2] member 4:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [3] member 5:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [4] member 6:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [5] member 7:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [6] member 8:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [7] member 9:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [8] member 10:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [9] member 11:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [10] member 12:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [11] member 13:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] TRANS [12] member 14:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [0] member 1:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (1.1193)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 5 high delivered 5 r=
eceived flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [1] member 2:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [2] member 3:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [3] member 4:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [4] member 5:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [5] member 6:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [6] member 7:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [7] member 8:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [8] member 9:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [9] member 10:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [10] member 11:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [11] member 12:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [12] member 13:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] position [13] member 14:
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] previous ringid (2.1192)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] aru 123 high delivered 1=
23 received flag 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] Did not need to originat=
e any messages in recovery.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 0, aru ffffffff
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 1, aru 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 2, aru 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] token retrans flag is 0 =
my set retrans flag0 retrans queue empty 1 count 3, aru 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] install seq 0 aru 0 high=
 seq received 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] retrans flag count 4 tok=
en aru 0 install seq 0 aru 0 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] Resetting old ring state
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] recovery to regular 1-0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] waiting_trans_ack change=
d to 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [MAIN  ] Member joined: r(0) ip(1=
0.3.94.89)=20
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [SYNC  ] call init for locally kn=
own services
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] entering OPERATIONAL sta=
te.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] A new membership (1.1197=
) was formed. Members joined: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [SYNC  ] enter sync process
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [SYNC  ] Committing synchronizati=
on for corosync configuration map access
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CMAP  ] Not first sync -> no act=
ion
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 3
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 4
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 5
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 6
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 7
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 8
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 9
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 10
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 11
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 12
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] downlist left_list: 0 re=
ceived
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got joinlist message fro=
m node 13
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [SYNC  ] Committing synchronizati=
on for corosync cluster closed process group service v1.01
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] my downlist: members(old=
:13 left:0)
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[0] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.110) , pid:30209
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[1] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.110) , pid:30209
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[2] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.109) , pid:31350
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[3] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.109) , pid:31350
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[4] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.108) , pid:3569
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[5] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.108) , pid:3569
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[6] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.107) , pid:19504
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[7] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.107) , pid:19504
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[8] gro=
up:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.97) , pid:11947
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[9] gro=
up:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.97) , pid:11947
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[10] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.96) , pid:20814
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[11] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.96) , pid:20814
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[12] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.95) , pid:39420
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[13] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.95) , pid:39420
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[14] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.94) , pid:12452
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[15] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.94) , pid:12452
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[16] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.93) , pid:44300
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[17] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.93) , pid:44300
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[18] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.92) , pid:42259
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[19] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.92) , pid:42259
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[20] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.91) , pid:40630
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[21] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.91) , pid:40630
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[22] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.90) , pid:25870
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[23] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.90) , pid:25870
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[24] gr=
oup:pve_kvstore_v1\x00, ip:r(0) ip(10.3.94.111) , pid:25634
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] joinlist_messages[25] gr=
oup:pve_dcdb_v1\x00, ip:r(0) ip(10.3.94.111) , pid:25634
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] Sending nodelist callbac=
k. ring_id =3D 1.1197
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 13
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[13]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 13
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[14]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[1]: vot=
es: 1, expected: 14 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: No Leavi=
ng: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVot=
e: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] total_votes=3D14, expect=
ed_votes=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 3 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 4 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 5 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 6 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 7 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 8 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 9 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 10 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 11 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 12 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 2 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] lowest node id: 1 us: 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] highest node id: 14 us: =
2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[2]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] total_votes=3D14, expect=
ed_votes=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 3 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 4 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 5 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 6 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 7 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 8 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 9 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 10 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 11 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 12 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 2 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] lowest node id: 1 us: 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] highest node id: 14 us: =
2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 3
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[3]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 3
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 4
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[4]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 4
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 5
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[5]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 5
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 6
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[6]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 6
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 7
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[7]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 7
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 8
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[8]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 8
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 9
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[9]: vot=
es: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 9
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 10
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[10]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 10
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 11
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[11]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 11
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 12
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[12]: vo=
tes: 1, expected: 14 flags: 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] flags: quorate: Yes Leav=
ing: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVo=
te: No QdeviceMasterWins: No
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] got nodeinfo message fro=
m cluster node 12
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] nodeinfo message[0]: vot=
es: 0, expected: 0 flags: 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [SYNC  ] Committing synchronizati=
on for corosync vote quorum service v1.0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] total_votes=3D14, expect=
ed_votes=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 1 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 3 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 4 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 5 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 6 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 7 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 8 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 9 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 10 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 11 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 12 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 13 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 14 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] node 2 state=3D1, votes=
=3D1, expected=3D14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] lowest node id: 1 us: 2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] highest node id: 14 us: =
2
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [QUORUM] Members[14]: 1 2 3 4 5 6=
 7 8 9 10 11 12 13 14
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [QUORUM] sending quorum notificat=
ion to (nil), length =3D 104
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [VOTEQ ] Sending quorum callback,=
 quorate =3D 1
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [MAIN  ] Completed service synchr=
onization, ready to provide service.
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [TOTEM ] waiting_trans_ack change=
d to 0
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got procjoin message fro=
m cluster node 1 (r(0) ip(10.3.94.89) ) for pid 16239
Sep 15 17:41:09 m6kvm2 corosync[25411]:   [CPG   ] got procjoin message fro=
m cluster node 1 (r(0) ip(10.3.94.89) ) for pid 16239

----- Mail original -----
De: "aderumier" <aderumier@odiso.com>
=C3=80: "Thomas Lamprecht" <t.lamprecht@proxmox.com>
Cc: "Proxmox VE development discussion" <pve-devel@lists.proxmox.com>
Envoy=C3=A9: Mardi 15 Septembre 2020 16:57:46
Objet: Re: [pve-devel] corosync bug: cluster break after 1 node clean shutd=
own

>>I mean this is bad, but also great!=20
>>Cam you do a coredump of the whole thing and upload it somewhere with the=
 version info=20
>>used (for dbgsym package)? That could help a lot.=20

I'll try to reproduce it again (with the full lock everywhere), and do the =
coredump.=20




I have tried the real time scheduling,=20

but I still have been able to reproduce the "lrm too long" for 60s (but as =
I'm restarting corosync each minute, I think it's unlocking=20
something at next corosync restart.)=20


this time it was blocked at the same time on a node in:=20

work {=20
...=20
} elsif ($state eq 'active') {=20
....=20
$self->update_lrm_status();=20


and another node in=20

if ($fence_request) {=20
$haenv->log('err', "node need to be fenced - releasing agent_lock\n");=20
$self->set_local_status({ state =3D> 'lost_agent_lock'});=20
} elsif (!$self->get_protected_ha_agent_lock()) {=20
$self->set_local_status({ state =3D> 'lost_agent_lock'});=20
} elsif ($self->{mode} eq 'maintenance') {=20
$self->set_local_status({ state =3D> 'maintenance'});=20
}=20





----- Mail original -----=20
De: "Thomas Lamprecht" <t.lamprecht@proxmox.com>=20
=C3=80: "aderumier" <aderumier@odiso.com>=20
Cc: "Proxmox VE development discussion" <pve-devel@lists.proxmox.com>=20
Envoy=C3=A9: Mardi 15 Septembre 2020 16:32:52=20
Objet: Re: [pve-devel] corosync bug: cluster break after 1 node clean shutd=
own=20

On 9/15/20 4:09 PM, Alexandre DERUMIER wrote:=20
>>> Can you try to give pmxcfs real time scheduling, e.g., by doing:=20
>>>=20
>>> # systemctl edit pve-cluster=20
>>>=20
>>> And then add snippet:=20
>>>=20
>>>=20
>>> [Service]=20
>>> CPUSchedulingPolicy=3Drr=20
>>> CPUSchedulingPriority=3D99=20
> yes, sure, I'll do it now=20
>=20
>=20
>> I'm currently digging the logs=20
>>> Is your most simplest/stable reproducer still a periodic restart of cor=
osync in one node?=20
> yes, a simple "systemctl restart corosync" on 1 node each minute=20
>=20
>=20
>=20
> After 1hour, it's still locked.=20
>=20
> on other nodes, I still have pmxfs logs like:=20
>=20

I mean this is bad, but also great!=20
Cam you do a coredump of the whole thing and upload it somewhere with the v=
ersion info=20
used (for dbgsym package)? That could help a lot.=20


> manual "pmxcfs -d"=20
> https://gist.github.com/aderumier/4cd91d17e1f8847b93ea5f621f257c2e=20
>=20

Hmm, the fuse connection of the previous one got into a weird state (or som=
ething is still=20
running) but I'd rather say this is a side-effect not directly connected to=
 the real bug.=20

>=20
> some interesting dmesg about "pvesr"=20
>=20
> [Tue Sep 15 14:45:34 2020] INFO: task pvesr:19038 blocked for more than 1=
20 seconds.=20
> [Tue Sep 15 14:45:34 2020] Tainted: P O 5.4.60-1-pve #1=20
> [Tue Sep 15 14:45:34 2020] "echo 0 > /proc/sys/kernel/hung_task_timeout_s=
ecs" disables this message.=20
> [Tue Sep 15 14:45:34 2020] pvesr D 0 19038 1 0x00000080=20
> [Tue Sep 15 14:45:34 2020] Call Trace:=20
> [Tue Sep 15 14:45:34 2020] __schedule+0x2e6/0x6f0=20
> [Tue Sep 15 14:45:34 2020] ? filename_parentat.isra.57.part.58+0xf7/0x180=
=20
> [Tue Sep 15 14:45:34 2020] schedule+0x33/0xa0=20
> [Tue Sep 15 14:45:34 2020] rwsem_down_write_slowpath+0x2ed/0x4a0=20
> [Tue Sep 15 14:45:34 2020] down_write+0x3d/0x40=20
> [Tue Sep 15 14:45:34 2020] filename_create+0x8e/0x180=20
> [Tue Sep 15 14:45:34 2020] do_mkdirat+0x59/0x110=20
> [Tue Sep 15 14:45:34 2020] __x64_sys_mkdir+0x1b/0x20=20
> [Tue Sep 15 14:45:34 2020] do_syscall_64+0x57/0x190=20
> [Tue Sep 15 14:45:34 2020] entry_SYSCALL_64_after_hwframe+0x44/0xa9=20
>=20

hmm, hangs in mkdir (cluster wide locking)=20


_______________________________________________=20
pve-devel mailing list=20
pve-devel@lists.proxmox.com=20
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel=20