public inbox for pbs-devel@lists.proxmox.com
 help / color / mirror / Atom feed
From: Christian Ebner <c.ebner@proxmox.com>
To: "Proxmox Backup Server development discussion"
	<pbs-devel@lists.proxmox.com>,
	"Fabian Grünbichler" <f.gruenbichler@proxmox.com>
Subject: Re: [pbs-devel] [PATCH proxmox-backup 1/3] GC: S3: reduce number of open FDs for to-be-deleted objects
Date: Fri, 21 Nov 2025 10:43:00 +0100	[thread overview]
Message-ID: <af638672-bbec-4a54-bb03-230fe9c83831@proxmox.com> (raw)
In-Reply-To: <20251121090605.262675-2-f.gruenbichler@proxmox.com>

one comment inline

On 11/21/25 10:06 AM, Fabian Grünbichler wrote:
> listing objects on the S3 side will return batches containing up to 1000
> objects. previously, if all those objects were garbage, phase2 would open and
> hold the lock file for each of them and delete them using a single call. this
> can easily run afoul the maximum number of open files allowed by the default
> process limits, which is 1024.
> 
> converting the code to instead delete batches of (at most) 100 objects should
> alleviate this issue until bumping the limit is deemed safe, while (in the
> worst case) causing 10x the number of delete requests.
> 
> Signed-off-by: Fabian Grünbichler <f.gruenbichler@proxmox.com>
> ---
>   pbs-datastore/src/datastore.rs | 25 ++++++++++++++++++++++++-
>   1 file changed, 24 insertions(+), 1 deletion(-)
> 
> diff --git a/pbs-datastore/src/datastore.rs b/pbs-datastore/src/datastore.rs
> index 0a5179230..1afcef53a 100644
> --- a/pbs-datastore/src/datastore.rs
> +++ b/pbs-datastore/src/datastore.rs
> @@ -58,6 +58,8 @@ pub const S3_DATASTORE_IN_USE_MARKER: &str = ".in-use";
>   const NAMESPACE_MARKER_FILENAME: &str = ".namespace";
>   // s3 put request times out after upload_size / 1 Kib/s, so about 2.3 hours for 8 MiB
>   const CHUNK_LOCK_TIMEOUT: Duration = Duration::from_secs(3 * 60 * 60);
> +// s3 deletion batch size to avoid 1024 open files soft limit
> +const S3_DELETE_BATCH_LIMIT: usize = 100;
>   
>   /// checks if auth_id is owner, or, if owner is a token, if
>   /// auth_id is the user of the token
> @@ -1657,7 +1659,7 @@ impl DataStore {
>                   proxmox_async::runtime::block_on(s3_client.list_objects_v2(&prefix, None))
>                       .context("failed to list chunk in s3 object store")?;
>   
> -            let mut delete_list = Vec::with_capacity(1000);
> +            let mut delete_list = Vec::with_capacity(S3_DELETE_BATCH_LIMIT);
>               loop {
>                   for content in list_bucket_result.contents {
>                       let (chunk_path, digest, bad) =
> @@ -1716,8 +1718,29 @@ impl DataStore {
>                       }
>   
>                       chunk_count += 1;
> +
> +                    // drop guard because of async S3 call below
> +                    drop(_guard);
> +
> +                    // limit pending deletes to avoid holding too many chunk flocks
> +                    if delete_list.len() > S3_DELETE_BATCH_LIMIT {

off by one error: this should never exceed the limit but rather delete 
once it is reached.

> +                        let delete_objects_result = proxmox_async::runtime::block_on(
> +                            s3_client.delete_objects(
> +                                &delete_list
> +                                    .iter()
> +                                    .map(|(key, _)| key.clone())
> +                                    .collect::<Vec<S3ObjectKey>>(),
> +                            ),
> +                        )?;
> +                        if let Some(_err) = delete_objects_result.error {
> +                            bail!("failed to delete some objects");
> +                        }
> +                        // release all chunk guards
> +                        delete_list.clear();
> +                    }
>                   }
>   
> +                // delete the last batch of objects, if there are any remaining
>                   if !delete_list.is_empty() {
>                       let delete_objects_result = proxmox_async::runtime::block_on(
>                           s3_client.delete_objects(



_______________________________________________
pbs-devel mailing list
pbs-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pbs-devel

  reply	other threads:[~2025-11-21  9:43 UTC|newest]

Thread overview: 10+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-11-21  9:05 [pbs-devel] [PATCH proxmox-backup 0/3] reduce GC S3 locking Fabian Grünbichler
2025-11-21  9:05 ` [pbs-devel] [PATCH proxmox-backup 1/3] GC: S3: reduce number of open FDs for to-be-deleted objects Fabian Grünbichler
2025-11-21  9:43   ` Christian Ebner [this message]
2025-11-21  9:06 ` [pbs-devel] [PATCH proxmox-backup 2/3] GC: S3: factor out batch object deletion Fabian Grünbichler
2025-11-21  9:06 ` [pbs-devel] [PATCH proxmox-backup 3/3] GC: S3: phase2: delete last partial batch of objects at the very end Fabian Grünbichler
2025-11-21  9:31   ` Christian Ebner
2025-11-21  9:46     ` Fabian Grünbichler
2025-11-21  9:53       ` Christian Ebner
2025-11-21 10:05 ` [pbs-devel] [PATCH proxmox-backup 0/3] reduce GC S3 locking Christian Ebner
2025-11-21 10:19 ` [pbs-devel] superseded: " Fabian Grünbichler

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=af638672-bbec-4a54-bb03-230fe9c83831@proxmox.com \
    --to=c.ebner@proxmox.com \
    --cc=f.gruenbichler@proxmox.com \
    --cc=pbs-devel@lists.proxmox.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal