From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from firstgate.proxmox.com (firstgate.proxmox.com [212.224.123.68]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by lists.proxmox.com (Postfix) with ESMTPS id B101070D2C for ; Wed, 15 Jun 2022 10:21:16 +0200 (CEST) Received: from firstgate.proxmox.com (localhost [127.0.0.1]) by firstgate.proxmox.com (Proxmox) with ESMTP id A773A276A6 for ; Wed, 15 Jun 2022 10:20:46 +0200 (CEST) Received: from proxmox-new.maurer-it.com (proxmox-new.maurer-it.com [94.136.29.106]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by firstgate.proxmox.com (Proxmox) with ESMTPS id B4AD627693 for ; Wed, 15 Jun 2022 10:20:45 +0200 (CEST) Received: from proxmox-new.maurer-it.com (localhost.localdomain [127.0.0.1]) by proxmox-new.maurer-it.com (Proxmox) with ESMTP id 8CEAD43B53 for ; Wed, 15 Jun 2022 10:20:45 +0200 (CEST) From: Stefan Sterz To: pbs-devel@lists.proxmox.com Date: Wed, 15 Jun 2022 10:20:38 +0200 Message-Id: <20220615082040.96959-3-s.sterz@proxmox.com> X-Mailer: git-send-email 2.30.2 In-Reply-To: <20220615082040.96959-1-s.sterz@proxmox.com> References: <20220615082040.96959-1-s.sterz@proxmox.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-SPAM-LEVEL: Spam detection results: 0 AWL -0.056 Adjusted score from AWL reputation of From: address BAYES_00 -1.9 Bayes spam probability is 0 to 1% KAM_DMARC_STATUS 0.01 Test Rule for DKIM or SPF Failure with Strict Alignment SPF_HELO_NONE 0.001 SPF: HELO does not publish an SPF Record SPF_PASS -0.001 SPF: sender matches SPF record T_SCC_BODY_TEXT_LINE -0.01 - URIBL_BLOCKED 0.001 ADMINISTRATOR NOTICE: The query to URIBL was blocked. See http://wiki.apache.org/spamassassin/DnsBlocklists#dnsbl-block for more information. [pull.rs] Subject: [pbs-devel] [RFC PATCH 2/4] fix #3786: server/datastore: add deep sync parameter to pull sync jobs X-BeenThere: pbs-devel@lists.proxmox.com X-Mailman-Version: 2.1.29 Precedence: list List-Id: Proxmox Backup Server development discussion List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 15 Jun 2022 08:21:16 -0000 Signed-off-by: Stefan Sterz --- pbs-datastore/src/backup_info.rs | 22 +++++++++++++++++++++- src/server/pull.rs | 28 ++++++++++++++++++---------- 2 files changed, 39 insertions(+), 11 deletions(-) diff --git a/pbs-datastore/src/backup_info.rs b/pbs-datastore/src/backup_info.rs index 10320a35..89461c66 100644 --- a/pbs-datastore/src/backup_info.rs +++ b/pbs-datastore/src/backup_info.rs @@ -9,7 +9,8 @@ use anyhow::{bail, format_err, Error}; use proxmox_sys::fs::{lock_dir_noblock, replace_file, CreateOptions}; use pbs_api_types::{ - Authid, BackupNamespace, BackupType, GroupFilter, BACKUP_DATE_REGEX, BACKUP_FILE_REGEX, + Authid, BackupNamespace, BackupType, GroupFilter, SnapshotVerifyState, VerifyState, + BACKUP_DATE_REGEX, BACKUP_FILE_REGEX, }; use pbs_config::{open_backup_lockfile, BackupLockGuard}; @@ -544,6 +545,25 @@ impl BackupDir { Ok(()) } + + /// Returns true if the last verification of the snapshot failed and false otherwise. + /// + /// Note that a snapshot that has not been verified will also return false. + pub fn is_corrupt(&self) -> bool { + let mut to_return = false; + + let _ = self.update_manifest(|m| { + let verify = m.unprotected["verify_state"].clone(); + + if let Ok(verify) = serde_json::from_value::(verify) { + if verify.state == VerifyState::Failed { + to_return = true; + } + } + }); + + to_return + } } impl AsRef for BackupDir { diff --git a/src/server/pull.rs b/src/server/pull.rs index 6778c66b..767b394c 100644 --- a/src/server/pull.rs +++ b/src/server/pull.rs @@ -57,7 +57,7 @@ pub struct PullParameters { /// How many levels of sub-namespaces to pull (0 == no recursion, None == maximum recursion) max_depth: Option, /// Whether to re-sync corrupted snapshots - _deep_sync: bool, + deep_sync: bool, /// Filters for reducing the pull scope group_filter: Option>, /// Rate limits for all transfers from `remote` @@ -111,7 +111,7 @@ impl PullParameters { owner, remove_vanished, max_depth, - _deep_sync: deep_sync, + deep_sync, group_filter, limit, }) @@ -371,6 +371,7 @@ async fn pull_snapshot( worker: &WorkerTask, reader: Arc, snapshot: &pbs_datastore::BackupDir, + params: &PullParameters, downloaded_chunks: Arc>>, ) -> Result<(), Error> { let mut manifest_name = snapshot.full_path(); @@ -437,7 +438,10 @@ async fn pull_snapshot( let mut path = snapshot.full_path(); path.push(&item.filename); - if path.exists() { + // if a snapshot could not be verified, the index file will stay the same, but it'll point + // to at least one corrupted chunk. hence, skip this check if the last verification job + // failed and we are running a deep sync. + if !(params.deep_sync && snapshot.is_corrupt()) && path.exists() { match archive_type(&item.filename)? { ArchiveType::DynamicIndex => { let index = DynamicIndexReader::open(&path)?; @@ -513,6 +517,7 @@ async fn pull_snapshot_from( worker: &WorkerTask, reader: Arc, snapshot: &pbs_datastore::BackupDir, + params: &PullParameters, downloaded_chunks: Arc>>, ) -> Result<(), Error> { let (_path, is_new, _snap_lock) = snapshot @@ -522,7 +527,7 @@ async fn pull_snapshot_from( if is_new { task_log!(worker, "sync snapshot {}", snapshot.dir()); - if let Err(err) = pull_snapshot(worker, reader, snapshot, downloaded_chunks).await { + if let Err(err) = pull_snapshot(worker, reader, snapshot, params, downloaded_chunks).await { if let Err(cleanup_err) = snapshot.datastore().remove_backup_dir( snapshot.backup_ns(), snapshot.as_ref(), @@ -535,7 +540,7 @@ async fn pull_snapshot_from( task_log!(worker, "sync snapshot {} done", snapshot.dir()); } else { task_log!(worker, "re-sync snapshot {}", snapshot.dir()); - pull_snapshot(worker, reader, snapshot, downloaded_chunks).await?; + pull_snapshot(worker, reader, snapshot, params, downloaded_chunks).await?; task_log!(worker, "re-sync snapshot {} done", snapshot.dir()); } @@ -666,10 +671,12 @@ async fn pull_group( remote_snapshots.insert(snapshot.time); - if let Some(last_sync_time) = last_sync { - if last_sync_time > snapshot.time { - skip_info.update(snapshot.time); - continue; + if !params.deep_sync { + if let Some(last_sync_time) = last_sync { + if last_sync_time > snapshot.time { + skip_info.update(snapshot.time); + continue; + } } } @@ -699,7 +706,8 @@ async fn pull_group( let snapshot = params.store.backup_dir(target_ns.clone(), snapshot)?; - let result = pull_snapshot_from(worker, reader, &snapshot, downloaded_chunks.clone()).await; + let result = + pull_snapshot_from(worker, reader, &snapshot, params, downloaded_chunks.clone()).await; progress.done_snapshots = pos as u64 + 1; task_log!(worker, "percentage done: {}", progress); -- 2.30.2