From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by lists.proxmox.com (Postfix) with ESMTPS id EB2E86CF98 for ; Thu, 12 Aug 2021 13:01:34 +0200 (CEST) Received: from firstgate.proxmox.com (localhost [127.0.0.1]) by firstgate.proxmox.com (Proxmox) with ESMTP id B361F2218E for ; Thu, 12 Aug 2021 13:01:34 +0200 (CEST) Received: from proxmox-new.maurer-it.com (proxmox-new.maurer-it.com [94.136.29.106]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by firstgate.proxmox.com (Proxmox) with ESMTPS id AE54021FF9 for ; Thu, 12 Aug 2021 13:01:24 +0200 (CEST) Received: from proxmox-new.maurer-it.com (localhost.localdomain [127.0.0.1]) by proxmox-new.maurer-it.com (Proxmox) with ESMTP id 90DF143315 for ; Thu, 12 Aug 2021 13:01:18 +0200 (CEST) From: Fabian Ebner To: pve-devel@lists.proxmox.com Date: Thu, 12 Aug 2021 13:01:11 +0200 Message-Id: <20210812110111.73883-13-f.ebner@proxmox.com> X-Mailer: git-send-email 2.30.2 In-Reply-To: <20210812110111.73883-1-f.ebner@proxmox.com> References: <20210812110111.73883-1-f.ebner@proxmox.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-SPAM-LEVEL: Spam detection results: 0 AWL 0.406 Adjusted score from AWL reputation of From: address BAYES_00 -1.9 Bayes spam probability is 0 to 1% KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment SPF_HELO_NONE 0.001 SPF: HELO does not publish an SPF Record SPF_PASS -0.001 SPF: sender matches SPF record Subject: [pve-devel] [RFC v3 guest-common 7/7] fix #3111: config: snapshot delete: check if replication still needs it X-BeenThere: pve-devel@lists.proxmox.com X-Mailman-Version: 2.1.29 Precedence: list List-Id: Proxmox VE development discussion List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 12 Aug 2021 11:01:35 -0000 and abort if it does and --force is not specified. After rollback, the rollback snapshot might still be needed as the base for incremental replication, because rollback removes (blocking) replication snapshots. It's not enough to limit the check to the most recent snapshot, because new snapshots might've been created between rollback and remove. It's not enough to limit the check to snapshots without a parent (i.e. in case of ZFS, the oldest), because some volumes might've been added only after that, meaning the oldest snapshot is not an incremental replication base for them. Signed-off-by: Fabian Ebner --- Sent as RFC, because I feel like this is quite a bit of code just to prevent a corner case that's now already warned about upon rollback. Arguably the warning in the UI is not very visible, but improving that by either using the new task warnings or showing the task viewer upon rollback is an alternative that might be preferable. src/PVE/AbstractConfig.pm | 41 ++++++++++++++++++++++++++++++++++++ src/PVE/Replication.pm | 6 +++++- src/PVE/ReplicationConfig.pm | 14 ++++++++++++ 3 files changed, 60 insertions(+), 1 deletion(-) diff --git a/src/PVE/AbstractConfig.pm b/src/PVE/AbstractConfig.pm index a5a15bf..39f1cc8 100644 --- a/src/PVE/AbstractConfig.pm +++ b/src/PVE/AbstractConfig.pm @@ -824,6 +824,44 @@ sub snapshot_create { $class->__snapshot_commit($vmid, $snapname); } +# Check if the snapshot might still be needed by a replication job. +my $snapshot_delete_assert_not_needed_by_replication = sub { + my ($class, $vmid, $conf, $snap, $snapname) = @_; + + my $repl_conf = PVE::ReplicationConfig->new(); + return if !$repl_conf->check_for_existing_jobs($vmid, 1); + + my $storecfg = PVE::Storage::config(); + + # Current config's volumes are relevant for replication. + my $volumes = $class->get_replicatable_volumes($storecfg, $vmid, $conf, 1); + + my $replication_jobs = $repl_conf->list_guests_local_replication_jobs($vmid); + + $class->foreach_volume($snap, sub { + my ($vs, $volume) = @_; + + my $volid_key = $class->volid_key(); + my $volid = $volume->{$volid_key}; + + return if !$volumes->{$volid}; + + my $snapshots = PVE::Storage::volume_snapshot_list($storecfg, $volid); + + for my $job ($replication_jobs->@*) { + my $jobid = $job->{id}; + + my @jobs_snapshots = grep { + PVE::Replication::is_replication_snapshot($_, $jobid) + } $snapshots->@*; + + next if scalar(@jobs_snapshots) > 0; + + die "snapshot '$snapname' needed by replication job '$jobid' - run replication first\n"; + } + }); +}; + # Deletes a snapshot. # Note: $drivehash is only set when called from snapshot_create. sub snapshot_delete { @@ -838,6 +876,9 @@ sub snapshot_delete { die "snapshot '$snapname' does not exist\n" if !defined($snap); + $snapshot_delete_assert_not_needed_by_replication->($class, $vmid, $conf, $snap, $snapname) + if !$drivehash && !$force; + $class->set_lock($vmid, 'snapshot-delete') if (!$drivehash); # doesn't already have a 'snapshot' lock diff --git a/src/PVE/Replication.pm b/src/PVE/Replication.pm index 2609ad6..098ac00 100644 --- a/src/PVE/Replication.pm +++ b/src/PVE/Replication.pm @@ -470,7 +470,11 @@ sub run_replication { } sub is_replication_snapshot { - my ($snapshot_name) = @_; + my ($snapshot_name, $jobid) = @_; + + if (defined($jobid)) { + return $snapshot_name =~ m/^__replicate_\Q$jobid\E/ ? 1 : 0; + } return $snapshot_name =~ m/^__replicate_/ ? 1 : 0; } diff --git a/src/PVE/ReplicationConfig.pm b/src/PVE/ReplicationConfig.pm index fd856a0..78f55bb 100644 --- a/src/PVE/ReplicationConfig.pm +++ b/src/PVE/ReplicationConfig.pm @@ -228,6 +228,20 @@ sub find_local_replication_job { return undef; } +sub list_guests_local_replication_jobs { + my ($cfg, $vmid) = @_; + + my $jobs = []; + + for my $job (values %{$cfg->{ids}}) { + next if $job->{type} ne 'local' || $job->{guest} != $vmid; + + push @{$jobs}, $job; + } + + return $jobs; +} + # makes old_target the new source for all local jobs of this guest # makes new_target the target for the single local job with target old_target sub switch_replication_job_target_nolock { -- 2.30.2