public inbox for pbs-devel@lists.proxmox.com
 help / color / mirror / Atom feed
From: Christian Ebner <c.ebner@proxmox.com>
To: pbs-devel@lists.proxmox.com
Subject: [pbs-devel] [PATCH v3 proxmox-backup 18/24] fix #3174: backup writer: inject queued chunk in upload steam
Date: Fri,  3 Nov 2023 14:37:23 +0100	[thread overview]
Message-ID: <20231103133729.2252730-19-c.ebner@proxmox.com> (raw)
In-Reply-To: <20231103133729.2252730-1-c.ebner@proxmox.com>

Inject the chunk in the backup writers upload stream, including them
thereby in the index file.

Signed-off-by: Christian Ebner <c.ebner@proxmox.com>
---
Changes since version 2:
- no changes

Changes since version 1:
- no changes

 pbs-client/src/backup_writer.rs | 84 +++++++++++++++++++--------------
 1 file changed, 49 insertions(+), 35 deletions(-)

diff --git a/pbs-client/src/backup_writer.rs b/pbs-client/src/backup_writer.rs
index cc6dd49a..0f18b1ff 100644
--- a/pbs-client/src/backup_writer.rs
+++ b/pbs-client/src/backup_writer.rs
@@ -23,7 +23,7 @@ use pbs_tools::crypt_config::CryptConfig;
 
 use proxmox_human_byte::HumanByte;
 
-use super::inject_reused_chunks::InjectChunks;
+use super::inject_reused_chunks::{InjectChunks, InjectReusedChunks, InjectedChunksInfo};
 use super::merge_known_chunks::{MergeKnownChunks, MergedChunkInfo};
 
 use super::{H2Client, HttpClient};
@@ -667,48 +667,62 @@ impl BackupWriter {
         let index_csum_2 = index_csum.clone();
 
         stream
-            .and_then(move |data| {
-                let chunk_len = data.len();
+            .inject_reused_chunks(injection_queue, stream_len, index_csum.clone())
+            .and_then(move |chunk_info| {
+                match chunk_info {
+                    InjectedChunksInfo::Known(chunks) => {
+                        total_chunks.fetch_add(chunks.len(), Ordering::SeqCst);
+                        future::ok(MergedChunkInfo::Known(chunks))
+                    }
+                    InjectedChunksInfo::Raw((offset, data)) => {
+                        let chunk_len = data.len();
 
-                total_chunks.fetch_add(1, Ordering::SeqCst);
-                let offset = stream_len.fetch_add(chunk_len, Ordering::SeqCst) as u64;
+                        total_chunks.fetch_add(1, Ordering::SeqCst);
 
-                let mut chunk_builder = DataChunkBuilder::new(data.as_ref()).compress(compress);
+                        let mut chunk_builder =
+                            DataChunkBuilder::new(data.as_ref()).compress(compress);
 
-                if let Some(ref crypt_config) = crypt_config {
-                    chunk_builder = chunk_builder.crypt_config(crypt_config);
-                }
+                        if let Some(ref crypt_config) = crypt_config {
+                            chunk_builder = chunk_builder.crypt_config(crypt_config);
+                        }
 
-                let mut known_chunks = known_chunks.lock().unwrap();
-                let digest = chunk_builder.digest();
+                        let mut known_chunks = known_chunks.lock().unwrap();
 
-                let mut guard = index_csum.lock().unwrap();
-                let csum = guard.as_mut().unwrap();
+                        let digest = chunk_builder.digest();
 
-                let chunk_end = offset + chunk_len as u64;
+                        let mut guard = index_csum.lock().unwrap();
+                        let csum = guard.as_mut().unwrap();
 
-                if !is_fixed_chunk_size {
-                    csum.update(&chunk_end.to_le_bytes());
-                }
-                csum.update(digest);
+                        let chunk_end = offset + chunk_len as u64;
 
-                let chunk_is_known = known_chunks.contains(digest);
-                if chunk_is_known {
-                    known_chunk_count.fetch_add(1, Ordering::SeqCst);
-                    reused_len.fetch_add(chunk_len, Ordering::SeqCst);
-                    future::ok(MergedChunkInfo::Known(vec![(offset, *digest)]))
-                } else {
-                    let compressed_stream_len2 = compressed_stream_len.clone();
-                    known_chunks.insert(*digest);
-                    future::ready(chunk_builder.build().map(move |(chunk, digest)| {
-                        compressed_stream_len2.fetch_add(chunk.raw_size(), Ordering::SeqCst);
-                        MergedChunkInfo::New(ChunkInfo {
-                            chunk,
-                            digest,
-                            chunk_len: chunk_len as u64,
-                            offset,
-                        })
-                    }))
+                        if !is_fixed_chunk_size {
+                            csum.update(&chunk_end.to_le_bytes());
+                        }
+                        csum.update(digest);
+
+                        let chunk_is_known = known_chunks.contains(digest);
+                        if chunk_is_known {
+                            known_chunk_count.fetch_add(1, Ordering::SeqCst);
+                            reused_len.fetch_add(chunk_len, Ordering::SeqCst);
+
+                            future::ok(MergedChunkInfo::Known(vec![(offset, *digest)]))
+                        } else {
+                            let compressed_stream_len2 = compressed_stream_len.clone();
+                            known_chunks.insert(*digest);
+
+                            future::ready(chunk_builder.build().map(move |(chunk, digest)| {
+                                compressed_stream_len2
+                                    .fetch_add(chunk.raw_size(), Ordering::SeqCst);
+
+                                MergedChunkInfo::New(ChunkInfo {
+                                    chunk,
+                                    digest,
+                                    chunk_len: chunk_len as u64,
+                                    offset,
+                                })
+                            }))
+                        }
+                    }
                 }
             })
             .merge_known_chunks()
-- 
2.39.2





  parent reply	other threads:[~2023-11-03 13:38 UTC|newest]

Thread overview: 25+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-11-03 13:37 [pbs-devel] [PATCH-SERIES v3 pxar proxmox-backup proxmox-widget-toolkit 00/24] fix #3174: improve file-level backup Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 1/24] fix #3174: decoder: factor out skip_bytes from skip_entry Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 2/24] fix #3174: decoder: impl skip_bytes for sync dec Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 3/24] fix #3174: encoder: calc filename + metadata byte size Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 4/24] fix #3174: enc/dec: impl PXAR_APPENDIX_REF entrytype Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 5/24] fix #3174: enc/dec: impl PXAR_APPENDIX entrytype Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 6/24] fix #3174: encoder: helper to add to encoder position Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 pxar 7/24] fix #3174: enc/dec: impl PXAR_APPENDIX_TAIL entrytype Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 08/24] fix #3174: index: add fn index list from start/end-offsets Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 09/24] fix #3174: index: add fn digest for DynamicEntry Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 10/24] fix #3174: api: double catalog upload size Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 11/24] fix #3174: catalog: introduce extended format v2 Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 12/24] fix #3174: archiver/extractor: impl appendix ref Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 13/24] fix #3174: catalog: add specialized Archive entry Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 14/24] fix #3174: extractor: impl seq restore from appendix Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 15/24] fix #3174: archiver: store ref to previous backup Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 16/24] fix #3174: upload stream: impl reused chunk injector Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 17/24] fix #3174: chunker: add forced boundaries Christian Ebner
2023-11-03 13:37 ` Christian Ebner [this message]
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 19/24] fix #3174: archiver: reuse files with unchanged metadata Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 20/24] fix #3174: specs: add backup detection mode specification Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 21/24] fix #3174: client: Add detection mode to backup creation Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 22/24] test-suite: add detection mode change benchmark Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-backup 23/24] test-suite: Add bin to deb, add shell completions Christian Ebner
2023-11-03 13:37 ` [pbs-devel] [PATCH v3 proxmox-widget-toolkit 24/24] file-browser: support pxar archive and fileref types Christian Ebner

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20231103133729.2252730-19-c.ebner@proxmox.com \
    --to=c.ebner@proxmox.com \
    --cc=pbs-devel@lists.proxmox.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal