From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by lists.proxmox.com (Postfix) with ESMTPS id 138F46CF94 for ; Thu, 12 Aug 2021 13:01:34 +0200 (CEST) Received: from firstgate.proxmox.com (localhost [127.0.0.1]) by firstgate.proxmox.com (Proxmox) with ESMTP id 0E2062212F for ; Thu, 12 Aug 2021 13:01:33 +0200 (CEST) Received: from proxmox-new.maurer-it.com (proxmox-new.maurer-it.com [94.136.29.106]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by firstgate.proxmox.com (Proxmox) with ESMTPS id 870EC21FF0 for ; Thu, 12 Aug 2021 13:01:24 +0200 (CEST) Received: from proxmox-new.maurer-it.com (localhost.localdomain [127.0.0.1]) by proxmox-new.maurer-it.com (Proxmox) with ESMTP id 4EB8D4331A for ; Thu, 12 Aug 2021 13:01:18 +0200 (CEST) From: Fabian Ebner To: pve-devel@lists.proxmox.com Date: Thu, 12 Aug 2021 13:01:05 +0200 Message-Id: <20210812110111.73883-7-f.ebner@proxmox.com> X-Mailer: git-send-email 2.30.2 In-Reply-To: <20210812110111.73883-1-f.ebner@proxmox.com> References: <20210812110111.73883-1-f.ebner@proxmox.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-SPAM-LEVEL: Spam detection results: 0 AWL 0.410 Adjusted score from AWL reputation of From: address BAYES_00 -1.9 Bayes spam probability is 0 to 1% KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment SPF_HELO_NONE 0.001 SPF: HELO does not publish an SPF Record SPF_PASS -0.001 SPF: sender matches SPF record Subject: [pve-devel] [PATCH v3 guest-common 1/7] partially fix #3111: snapshot rollback: improve removing replication snapshots X-BeenThere: pve-devel@lists.proxmox.com X-Mailman-Version: 2.1.29 Precedence: list List-Id: Proxmox VE development discussion List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 12 Aug 2021 11:01:34 -0000 Get the replicatable volumes from the snapshot config rather than the current config. And filter those volumes further to those that will actually be rolled back. Previously, a volume that only had replication snapshots (e.g. because it was added after the snapshot was taken, or the vmstate volume) would lose them. Then, on the next replication run, such a volume would lead to an error, because replication tried to do a full sync, but the target volume still exists. This is not a complete fix. It is still possible to run into problems: - by removing the last (non-replication) snapshots after a rollback before replication can run once. - by creating a snapshot and making a rollback before replication can run once. The list of volumes is not required to be sorted for prepare(), but it is sorted by how foreach_volume() iterates now, so not random. Signed-off-by: Fabian Ebner --- Changes from v2: * mention other possible way to run into issue in commit message * remove redundant comment src/PVE/AbstractConfig.pm | 17 +++++++++++++---- 1 file changed, 13 insertions(+), 4 deletions(-) diff --git a/src/PVE/AbstractConfig.pm b/src/PVE/AbstractConfig.pm index 3348d8a..493bf97 100644 --- a/src/PVE/AbstractConfig.pm +++ b/src/PVE/AbstractConfig.pm @@ -974,13 +974,22 @@ sub snapshot_rollback { if ($prepare) { my $repl_conf = PVE::ReplicationConfig->new(); if ($repl_conf->check_for_existing_jobs($vmid, 1)) { - # remove all replication snapshots - my $volumes = $class->get_replicatable_volumes($storecfg, $vmid, $conf, 1); - my $sorted_volids = [ sort keys %$volumes ]; + my $volumes = $class->get_replicatable_volumes($storecfg, $vmid, $snap, 1); + + # filter by what we actually iterate over below (excludes vmstate!) + my $volids = []; + $class->foreach_volume($snap, sub { + my ($vs, $volume) = @_; + + my $volid_key = $class->volid_key(); + my $volid = $volume->{$volid_key}; + + push @{$volids}, $volid if $volumes->{$volid}; + }); # remove all local replication snapshots (jobid => undef) my $logfunc = sub { my $line = shift; chomp $line; print "$line\n"; }; - PVE::Replication::prepare($storecfg, $sorted_volids, undef, 1, undef, $logfunc); + PVE::Replication::prepare($storecfg, $volids, undef, 1, undef, $logfunc); } $class->foreach_volume($snap, sub { -- 2.30.2