* [pbs-devel] Bulk initial sync from remote
@ 2021-03-26 15:15 Sebastian
0 siblings, 0 replies; 2+ messages in thread
From: Sebastian @ 2021-03-26 15:15 UTC (permalink / raw)
To: pbs-devel
Good afternoon everyone,
is it possible to do an initial bulk sync from a remote? (using external storage media for example)
E.g. can all files (chunk directory etc.) be blindy copied from one pbs server to the remote pbs server using a external storage medium?
Use-case: doing a initial full sync from a remote can cost a lot of bandwidth (or time), while incrementals can be small (when aren't a lot of changes).
Thanks in advance.
Cheers,
Sebastian
^ permalink raw reply [flat|nested] 2+ messages in thread
* Re: [pbs-devel] Bulk initial sync from remote
@ 2021-03-26 16:24 Fabian Grünbichler
0 siblings, 0 replies; 2+ messages in thread
From: Fabian Grünbichler @ 2021-03-26 16:24 UTC (permalink / raw)
To: Proxmox Backup Server development discussion, Sebastian
> Sebastian <s.schauenburg@gmail.com> hat am 26.03.2021 16:15 geschrieben:
>
> Good afternoon everyone,
>
> is it possible to do an initial bulk sync from a remote? (using external storage media for example)
> E.g. can all files (chunk directory etc.) be blindy copied from one pbs server to the remote pbs server using a external storage medium?
yes. a "blind copy" does risk a certain amount of inconsistency if there are any concurrent actions on the datastore (e.g., if you first copy all the snapshot metadata first, then continue with .chunks, and now a prune + GC run happens and deletes some chunks that you haven't copied yet).
you can avoid that by:
- defining the external medium as datastore, configure a 'local' remote pointing to the same node, and use the sync/pull mechanism instead of a blind copy (that will iterate over snapshots and copy associated chunks together with the snapshot metadata, so you'll never copy orphaned chunks or snapshot metadata without associated chunks). this will incur network/TLS overhead since it works over the API
- do a two-phase rsync or similar, and ensure the datastore is quiet for the final (small) sync
after moving your external disk, you need to manually create the datastore.cfg entry (or create a datastore using the GUI with a different path, and then edit it to point it to your actual path, or copy the contents from your external media into the created directory).
a datastore directory with the .chunks subdir and the backup type directories (by default: vm, ct, host) is self-contained as far as stored backups are concerned. scheduled jobs (prune, verify, GC) are stored outside, so those need to be recreated if you just have the "raw" datastore.
> Use-case: doing a initial full sync from a remote can cost a lot of bandwidth (or time), while incrementals can be small (when aren't a lot of changes).
common use case, should work with the caveats noted above :)
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2021-03-26 16:24 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-03-26 15:15 [pbs-devel] Bulk initial sync from remote Sebastian
2021-03-26 16:24 Fabian Grünbichler
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox