From: Christian Ebner <c.ebner@proxmox.com>
To: Proxmox Backup Server development discussion
<pbs-devel@lists.proxmox.com>, Nicolas Frey <n.frey@proxmox.com>
Subject: Re: [pbs-devel] [PATCH proxmox-backup 2/4] api: verify: use worker-threads to determine the number of threads to use
Date: Thu, 6 Nov 2025 10:09:15 +0100 [thread overview]
Message-ID: <b3dd6f0c-807e-442c-94bc-d34deaf62882@proxmox.com> (raw)
In-Reply-To: <20251105155129.517430-4-n.frey@proxmox.com>
Please add a short commit message describing what the worker threads
cover, e.g. that this parameter controls the number of reader and chunk
verification threads.
What tripped me over just now:
Is this intentionally not increasing the number of chunk verification
threads? Or was that overlooked? From the name of the parameter I
suspected this to act on both, reading and verifying. If this is not the
case, maybe the parameter should get renamed to a more telling
`parallel-chunk-readers` instead?
further comment inline
On 11/5/25 4:51 PM, Nicolas Frey wrote:
> Signed-off-by: Nicolas Frey <n.frey@proxmox.com>
> ---
> src/api2/admin/datastore.rs | 13 +++++++++++--
> src/api2/backup/environment.rs | 2 +-
> src/backup/verify.rs | 5 ++++-
> src/server/verify_job.rs | 3 ++-
> 4 files changed, 18 insertions(+), 5 deletions(-)
>
> diff --git a/src/api2/admin/datastore.rs b/src/api2/admin/datastore.rs
> index d192ee39..69a09081 100644
> --- a/src/api2/admin/datastore.rs
> +++ b/src/api2/admin/datastore.rs
> @@ -677,6 +677,14 @@ pub async fn status(
> schema: NS_MAX_DEPTH_SCHEMA,
> optional: true,
> },
> + "worker-threads": {
> + description: "Set the number of worker threads to use for the job",
> + type: Integer,
> + optional: true,
> + minimum: 1,
> + maximum: 32,
> + default: 1,
> + },
As mentioned on the pbs-api-types patch, this should reuse the same
schema as (will be) defined there, so this does not be to be re-defined
and stays in sync.
> },
> },
> returns: {
> @@ -690,7 +698,7 @@ pub async fn status(
> )]
> /// Verify backups.
> ///
> -/// This function can verify a single backup snapshot, all backup from a backup group,
> +/// This function can verify a single backup snapshot, all backups from a backup group,
> /// or all backups in the datastore.
> #[allow(clippy::too_many_arguments)]
> pub fn verify(
> @@ -702,6 +710,7 @@ pub fn verify(
> ignore_verified: Option<bool>,
> outdated_after: Option<i64>,
> max_depth: Option<usize>,
> + worker_threads: Option<usize>,
this could be a plain `usize` already, so it does not need to be
unwrapped for each parallel worker instantiation. The unwrapping and
setting to default can already happen in the constructor.
> rpcenv: &mut dyn RpcEnvironment,
> ) -> Result<Value, Error> {
> let auth_id: Authid = rpcenv.get_auth_id().unwrap().parse()?;
> @@ -781,7 +790,7 @@ pub fn verify(
> auth_id.to_string(),
> to_stdout,
> move |worker| {
> - let verify_worker = VerifyWorker::new(worker.clone(), datastore)?;
> + let verify_worker = VerifyWorker::new(worker.clone(), datastore, worker_threads)?;
> let failed_dirs = if let Some(backup_dir) = backup_dir {
> let mut res = Vec::new();
> if !verify_worker.verify_backup_dir(
> diff --git a/src/api2/backup/environment.rs b/src/api2/backup/environment.rs
> index 0e8eab1b..5e6a73b9 100644
> --- a/src/api2/backup/environment.rs
> +++ b/src/api2/backup/environment.rs
> @@ -812,7 +812,7 @@ impl BackupEnvironment {
> move |worker| {
> worker.log_message("Automatically verifying newly added snapshot");
>
> - let verify_worker = VerifyWorker::new(worker.clone(), datastore)?;
> + let verify_worker = VerifyWorker::new(worker.clone(), datastore, None)?;
> if !verify_worker.verify_backup_dir_with_lock(
> &backup_dir,
> worker.upid().clone(),
> diff --git a/src/backup/verify.rs b/src/backup/verify.rs
> index 7f91f38c..e11dba8e 100644
> --- a/src/backup/verify.rs
> +++ b/src/backup/verify.rs
> @@ -32,6 +32,7 @@ pub struct VerifyWorker {
> verified_chunks: Arc<Mutex<HashSet<[u8; 32]>>>,
> corrupt_chunks: Arc<Mutex<HashSet<[u8; 32]>>>,
> backend: DatastoreBackend,
> + worker_threads: Option<usize>,
... plain `usize` here
> }
>
> impl VerifyWorker {
> @@ -39,6 +40,7 @@ impl VerifyWorker {
> pub fn new(
> worker: Arc<dyn WorkerTaskContext>,
> datastore: Arc<DataStore>,
> + worker_threads: Option<usize>,
> ) -> Result<Self, Error> {
> let backend = datastore.backend()?;
> Ok(Self {
> @@ -49,6 +51,7 @@ impl VerifyWorker {
> // start with 64 chunks since we assume there are few corrupt ones
> corrupt_chunks: Arc::new(Mutex::new(HashSet::with_capacity(64))),
> backend,
> + worker_threads,
unwrap_or(4) here... or even define a constant for the default value,
although if it is placed here, it will only occur once.
> })
> }
>
> @@ -220,7 +223,7 @@ impl VerifyWorker {
> .datastore
> .get_chunks_in_order(&*index, skip_chunk, check_abort)?;
>
> - let reader_pool = ParallelHandler::new("read chunks", 4, {
> + let reader_pool = ParallelHandler::new("read chunks", self.worker_threads.unwrap_or(4), {
> let decoder_pool = decoder_pool.channel();
> let datastore = Arc::clone(&self.datastore);
> let corrupt_chunks = Arc::clone(&self.corrupt_chunks);
> diff --git a/src/server/verify_job.rs b/src/server/verify_job.rs
> index c8792174..9d790b07 100644
> --- a/src/server/verify_job.rs
> +++ b/src/server/verify_job.rs
> @@ -41,7 +41,8 @@ pub fn do_verification_job(
> None => Default::default(),
> };
>
> - let verify_worker = VerifyWorker::new(worker.clone(), datastore)?;
> + let verify_worker =
> + VerifyWorker::new(worker.clone(), datastore, verification_job.worker_threads)?;
> let result = verify_worker.verify_all_backups(
> worker.upid(),
> ns,
_______________________________________________
pbs-devel mailing list
pbs-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pbs-devel
next prev parent reply other threads:[~2025-11-06 9:09 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-05 15:51 [pbs-devel] [PATCH proxmox{, -backup} 0/5] parallelize chunk reads in verification Nicolas Frey
2025-11-05 15:51 ` [pbs-devel] [PATCH proxmox 1/1] pbs-api-types: jobs: verify: add worker-threads to VerificationJobConfig Nicolas Frey
2025-11-06 8:14 ` Christian Ebner
2025-11-05 15:51 ` [pbs-devel] [PATCH proxmox-backup 1/4] api: verify: move chunk loading into parallel handler Nicolas Frey
2025-11-06 8:54 ` Christian Ebner
2025-11-06 9:04 ` Nicolas Frey
2025-11-06 9:26 ` Christian Ebner
2025-11-05 15:51 ` [pbs-devel] [PATCH proxmox-backup 2/4] api: verify: use worker-threads to determine the number of threads to use Nicolas Frey
2025-11-06 9:09 ` Christian Ebner [this message]
2025-11-06 9:23 ` Nicolas Frey
2025-11-06 9:32 ` Christian Ebner
2025-11-06 11:22 ` Nicolas Frey
2025-11-06 11:57 ` Christian Ebner
2025-11-05 15:51 ` [pbs-devel] [PATCH proxmox-backup 3/4] api: verify: add worker-threads to update endpoint Nicolas Frey
2025-11-06 9:13 ` Christian Ebner
2025-11-05 15:51 ` [pbs-devel] [PATCH proxmox-backup 4/4] ui: verify: add option to set number of threads for job Nicolas Frey
2025-11-06 9:22 ` Christian Ebner
2025-11-06 9:25 ` Nicolas Frey
2025-11-06 8:02 ` [pbs-devel] [PATCH proxmox{, -backup} 0/5] parallelize chunk reads in verification Christian Ebner
2025-11-06 16:15 ` [pbs-devel] superseded: " Nicolas Frey
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=b3dd6f0c-807e-442c-94bc-d34deaf62882@proxmox.com \
--to=c.ebner@proxmox.com \
--cc=n.frey@proxmox.com \
--cc=pbs-devel@lists.proxmox.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.