From: Christian Ebner <c.ebner@proxmox.com>
To: pbs-devel@lists.proxmox.com
Subject: [pbs-devel] [PATCH proxmox-backup v10 40/46] datastore: conditionally upload atime marker chunk to s3 backend
Date: Mon, 21 Jul 2025 18:45:01 +0200 [thread overview]
Message-ID: <20250721164507.1045869-44-c.ebner@proxmox.com> (raw)
In-Reply-To: <20250721164507.1045869-1-c.ebner@proxmox.com>
Since commit b18eab64 ("fix #5982: garbage collection: check atime
updates are honored"), the 4 MiB fixed sized, unencypted and
compressed chunk containing all zeros is inserted at datastore
creation if the atime safety check is enabled.
If the datastore is backed by an S3 object store, chunk uploads are
avoided by checking the presence of the chunks in the local cache
store. Therefore, the all zero chunk will however not be uploaded
since already inserted locally.
Fix this by conditionally uploading the chunk before performing the
atime update check for datastores backed by S3.
Signed-off-by: Christian Ebner <c.ebner@proxmox.com>
---
changes since version 9:
- no changes
pbs-datastore/src/chunk_store.rs | 25 ++++++++++++++++++++++---
pbs-datastore/src/datastore.rs | 20 ++++++++++----------
src/api2/config/datastore.rs | 5 ++++-
3 files changed, 36 insertions(+), 14 deletions(-)
diff --git a/pbs-datastore/src/chunk_store.rs b/pbs-datastore/src/chunk_store.rs
index 95f00e8d5..3c59612bb 100644
--- a/pbs-datastore/src/chunk_store.rs
+++ b/pbs-datastore/src/chunk_store.rs
@@ -9,6 +9,7 @@ use tracing::{info, warn};
use pbs_api_types::{DatastoreFSyncLevel, GarbageCollectionStatus};
use proxmox_io::ReadExt;
+use proxmox_s3_client::S3Client;
use proxmox_sys::fs::{create_dir, create_path, file_type_from_file_stat, CreateOptions};
use proxmox_sys::process_locker::{
ProcessLockExclusiveGuard, ProcessLockSharedGuard, ProcessLocker,
@@ -454,11 +455,29 @@ impl ChunkStore {
/// Uses a 4 MiB fixed size, compressed but unencrypted chunk to test. The chunk is inserted in
/// the chunk store if not yet present.
/// Returns with error if the check could not be performed.
- pub fn check_fs_atime_updates(&self, retry_on_file_changed: bool) -> Result<(), Error> {
+ pub fn check_fs_atime_updates(
+ &self,
+ retry_on_file_changed: bool,
+ s3_client: Option<Arc<S3Client>>,
+ ) -> Result<(), Error> {
let (zero_chunk, digest) = DataChunkBuilder::build_zero_chunk(None, 4096 * 1024, true)?;
- let (pre_existing, _) = self.insert_chunk(&zero_chunk, &digest)?;
let (path, _digest) = self.chunk_path(&digest);
+ if let Some(ref s3_client) = s3_client {
+ if let Err(err) = std::fs::metadata(&path) {
+ if err.kind() == std::io::ErrorKind::NotFound {
+ let object_key = crate::s3::object_key_from_digest(&digest)?;
+ proxmox_async::runtime::block_on(s3_client.upload_no_replace_with_retry(
+ object_key,
+ zero_chunk.raw_data().to_vec().into(),
+ ))
+ .context("failed to upload chunk to s3 backend")?;
+ }
+ }
+ }
+
+ let (pre_existing, _) = self.insert_chunk(&zero_chunk, &digest)?;
+
// Take into account timestamp update granularity in the kernel
// Blocking the thread is fine here since this runs in a worker.
std::thread::sleep(Duration::from_secs(1));
@@ -478,7 +497,7 @@ impl ChunkStore {
// two metadata calls, try to check once again on changed file
if metadata_before.ino() != metadata_now.ino() {
if retry_on_file_changed {
- return self.check_fs_atime_updates(false);
+ return self.check_fs_atime_updates(false, s3_client);
}
bail!("chunk {path:?} changed twice during access time safety check, cannot proceed.");
}
diff --git a/pbs-datastore/src/datastore.rs b/pbs-datastore/src/datastore.rs
index e6dfa8ee7..4526bf2b8 100644
--- a/pbs-datastore/src/datastore.rs
+++ b/pbs-datastore/src/datastore.rs
@@ -1543,10 +1543,19 @@ impl DataStore {
.parse_property_string(gc_store_config.tuning.as_deref().unwrap_or(""))?,
)?;
+ let s3_client = match self.backend()? {
+ DatastoreBackend::Filesystem => None,
+ DatastoreBackend::S3(s3_client) => {
+ proxmox_async::runtime::block_on(s3_client.head_bucket())
+ .context("failed to reach bucket")?;
+ Some(s3_client)
+ }
+ };
+
if tuning.gc_atime_safety_check.unwrap_or(true) {
self.inner
.chunk_store
- .check_fs_atime_updates(true)
+ .check_fs_atime_updates(true, s3_client.clone())
.context("atime safety check failed")?;
info!("Access time update check successful, proceeding with GC.");
} else {
@@ -1585,15 +1594,6 @@ impl DataStore {
1024 * 1024
};
- let s3_client = match self.backend()? {
- DatastoreBackend::Filesystem => None,
- DatastoreBackend::S3(s3_client) => {
- proxmox_async::runtime::block_on(s3_client.head_bucket())
- .context("failed to reach bucket")?;
- Some(s3_client)
- }
- };
-
info!("Start GC phase1 (mark used chunks)");
self.mark_used_chunks(
diff --git a/src/api2/config/datastore.rs b/src/api2/config/datastore.rs
index 7a4a39074..9b87e01f2 100644
--- a/src/api2/config/datastore.rs
+++ b/src/api2/config/datastore.rs
@@ -1,4 +1,5 @@
use std::path::{Path, PathBuf};
+use std::sync::Arc;
use ::serde::{Deserialize, Serialize};
use anyhow::{bail, format_err, Context, Error};
@@ -118,6 +119,7 @@ pub(crate) fn do_create_datastore(
.parse_property_string(datastore.tuning.as_deref().unwrap_or(""))?,
)?;
+ let mut backend_s3_client = None;
if let Some(ref backend_config) = datastore.backend {
let backend_config: DatastoreBackendConfig = backend_config.parse()?;
match backend_config.ty.unwrap_or_default() {
@@ -151,6 +153,7 @@ pub(crate) fn do_create_datastore(
// Fine to block since this runs in worker task
proxmox_async::runtime::block_on(s3_client.head_bucket())
.context("failed to access bucket")?;
+ backend_s3_client = Some(Arc::new(s3_client));
}
}
}
@@ -194,7 +197,7 @@ pub(crate) fn do_create_datastore(
if tuning.gc_atime_safety_check.unwrap_or(true) {
chunk_store
- .check_fs_atime_updates(true)
+ .check_fs_atime_updates(true, backend_s3_client)
.context("access time safety check failed")?;
info!("Access time update check successful.");
} else {
--
2.47.2
_______________________________________________
pbs-devel mailing list
pbs-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pbs-devel
next prev parent reply other threads:[~2025-07-21 16:48 UTC|newest]
Thread overview: 56+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-07-21 16:44 [pbs-devel] [PATCH proxmox{, -backup} v10 00/49] fix #2943: S3 storage backend for datastores Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox v10 1/3] pbs-api-types: extend datastore config by backend config enum Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox v10 2/3] pbs-api-types: maintenance: add new maintenance mode S3 refresh Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox v10 3/3] s3 client: Add missing S3 object key max length check Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 01/46] datastore: add helpers for path/digest to s3 object key conversion Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 02/46] config: introduce s3 object store client configuration Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 03/46] api: config: implement endpoints to manipulate and list s3 configs Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 04/46] api: datastore: check s3 backend bucket access on datastore create Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 05/46] api/cli: add endpoint and command to check s3 client connection Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 06/46] datastore: allow to get the backend for a datastore Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 07/46] api: backup: store datastore backend in runtime environment Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 08/46] api: backup: conditionally upload chunks to s3 object store backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 09/46] api: backup: conditionally upload blobs " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 10/46] api: backup: conditionally upload indices " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 11/46] api: backup: conditionally upload manifest " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 12/46] api: datastore: conditionally upload client log to s3 backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 13/46] sync: pull: conditionally upload content " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 14/46] api: reader: fetch chunks based on datastore backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 15/46] datastore: local chunk reader: read chunks based on backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 16/46] verify worker: add datastore backed to verify worker Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 17/46] verify: implement chunk verification for stores with s3 backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 18/46] datastore: create namespace marker in " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 19/46] datastore: create/delete protected marker file on s3 storage backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 20/46] datastore: prune groups/snapshots from s3 object store backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 21/46] datastore: get and set owner for s3 " Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 22/46] datastore: implement garbage collection for s3 backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 23/46] ui: add datastore type selector and reorganize component layout Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 24/46] ui: add s3 client edit window for configuration create/edit Christian Ebner
2025-07-21 20:14 ` Thomas Lamprecht
2025-07-22 6:24 ` Christian Ebner
2025-07-22 7:00 ` Thomas Lamprecht
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 25/46] ui: add s3 client view for configuration Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 26/46] ui: expose the s3 client view in the navigation tree Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 27/46] ui: add s3 client selector and bucket field for s3 backend setup Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 28/46] tools: lru cache: add removed callback for evicted cache nodes Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 29/46] tools: async lru cache: implement insert, remove and contains methods Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 30/46] datastore: add local datastore cache for network attached storages Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 31/46] api: backup: use local datastore cache on s3 backend chunk upload Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 32/46] api: reader: use local datastore cache on s3 backend chunk fetching Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 33/46] datastore: local chunk reader: get cached chunk from local cache store Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 34/46] backup writer: refactor parameters into backup writer options struct Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 35/46] api: backup: add no-cache flag to bypass local datastore cache Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 36/46] api/datastore: implement refresh endpoint for stores with s3 backend Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 37/46] cli: add dedicated subcommand for datastore s3 refresh Christian Ebner
2025-07-21 16:44 ` [pbs-devel] [PATCH proxmox-backup v10 38/46] ui: render s3 refresh as valid maintenance type and task description Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 39/46] ui: expose s3 refresh button for datastores backed by object store Christian Ebner
2025-07-21 16:45 ` Christian Ebner [this message]
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 41/46] bin: implement client subcommands for s3 configuration manipulation Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 42/46] bin: expose reuse-datastore flag for proxmox-backup-manager Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 43/46] datastore: mark store as in-use by setting marker on s3 backend Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 44/46] datastore: run s3-refresh when reusing a datastore with " Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 45/46] api/ui: add flag to allow overwriting in-use marker for " Christian Ebner
2025-07-21 16:45 ` [pbs-devel] [PATCH proxmox-backup v10 46/46] docs: Add section describing how to setup s3 backed datastore Christian Ebner
2025-07-22 8:14 ` [pbs-devel] [PATCH proxmox{, -backup} v10 00/49] fix #2943: S3 storage backend for datastores Hannes Laimer
2025-07-22 9:29 ` Lukas Wagner
2025-07-22 10:13 ` [pbs-devel] superseded: " Christian Ebner
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250721164507.1045869-44-c.ebner@proxmox.com \
--to=c.ebner@proxmox.com \
--cc=pbs-devel@lists.proxmox.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox