* Re: [PVE-User] proxmox-restore - performance issues
@ 2021-10-01 6:52 Gregor Burck
2021-10-01 7:00 ` Dominik Csapak
0 siblings, 1 reply; 8+ messages in thread
From: Gregor Burck @ 2021-10-01 6:52 UTC (permalink / raw)
To: Proxmox VE user list
Hi,
thank you for reply. I made a lot of different tests and setups, but
this the setup I want to use:
Original setup:
HP DL380 Gen9 with
E5-2640 v3 @ 2.60GHz
256 GB RAM
2x SSDs for host OS
For an ZFS Rais 10:
2x 1TB SAMSUNG NVME PM983 for spezial devices
12x 8 TB HP SAS HDDs
root@ph-pbs:~# zpool status
pool: ZFSPOOL
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
ZFSPOOL ONLINE 0 0 0
mirror-0 ONLINE 0 0 0
sdc ONLINE 0 0 0
sdd ONLINE 0 0 0
mirror-1 ONLINE 0 0 0
sde ONLINE 0 0 0
sdf ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
sdg ONLINE 0 0 0
sdh ONLINE 0 0 0
mirror-3 ONLINE 0 0 0
sdi ONLINE 0 0 0
sdj ONLINE 0 0 0
mirror-4 ONLINE 0 0 0
sdk ONLINE 0 0 0
sdl ONLINE 0 0 0
mirror-5 ONLINE 0 0 0
sdm ONLINE 0 0 0
sdn ONLINE 0 0 0
special
mirror-6 ONLINE 0 0 0
nvme0n1 ONLINE 0 0 0
nvme1n1 ONLINE 0 0 0
errors: No known data errors
pool: rpool
state: ONLINE
scan: scrub repaired 0B in 00:02:40 with 0 errors on Sun Aug 8
00:26:43 2021
config:
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
mirror-0 ONLINE 0 0 0
sda3 ONLINE 0 0 0
sdb3 ONLINE 0 0 0
errors: No known data errors
The VMSTORE and the BACKUPSTORE is on the zsf as an dataset:
root@ph-pbs:~# zfs list
NAME USED AVAIL REFER MOUNTPOINT
ZFSPOOL 10.1T 32.1T 96K /ZFSPOOL
ZFSPOOL/BACKUPSTORE001 5.63T 32.1T 5.63T /ZFSPOOL/BACKUPSTORE001
ZFSPOOL/VMSTORE001 4.52T 32.1T 4.52T /ZFSPOOL/VMSTORE001
rpool 27.3G 80.2G 96K /rpool
rpool/ROOT 27.3G 80.2G 96K /rpool/ROOT
rpool/ROOT/pbs-1 27.3G 80.2G 27.3G /
The VM I tested with is our Exchange Server. Raw image size 500GB,
netto ~400GB content
First Test with one restore job:
Virtual
Environment 7.0-11
Datacenter
Search:
Logs
new
volume ID is 'VMSTORE:vm-101-disk-0'
restore
proxmox backup image: /usr/bin/pbs-restore --repository
root@pam@ph-pbs.peiker-holding.de:ZFSPOOLBACKUP
vm/121/2021-07-23T19:00:03Z drive-virtio0.img.fidx
/dev/zvol/ZFSPOOLVMSTORE/vm-101-disk-0 --verbose --format raw
--skip-zero
connecting
to repository 'root@pam@ph-pbs.peiker-holding.de:ZFSPOOLBACKUP'
open
block backend for target '/dev/zvol/ZFSPOOLVMSTORE/vm-101-disk-0'
starting
to restore snapshot 'vm/121/2021-07-23T19:00:03Z'
download
and verify backup index
progress
1% (read 5368709120 bytes, zeroes = 2% (125829120 bytes), duration 86
sec)
progress
2% (read 10737418240 bytes, zeroes = 1% (159383552 bytes), duration
181 sec)
progress
3% (read 16106127360 bytes, zeroes = 0% (159383552 bytes), duration
270 sec)
.
.
progress
98% (read 526133493760 bytes, zeroes = 0% (3628072960 bytes),
duration 9492 sec)
progress
99% (read 531502202880 bytes, zeroes = 0% (3628072960 bytes),
duration 9583 sec)
progress
100% (read 536870912000 bytes, zeroes = 0% (3628072960 bytes),
duration 9676 sec)
restore
image complete (bytes=536870912000, duration=9676.97s,
speed=52.91MB/s)
rescan
volumes...
TASK
OK
When I regard iotop I see about the same rate.
But when I start multiple restore jobs parallel, I see that the single
jon is still on IO 40-50 MB/s but the total IO is multiple of the
rate. I see on iotop rates to 200-250 MB/s
So I guess it isn't the store. In some Test with an Setup where I used
the nvmes as source and target I could reach a singele restore rate
about 70 MB/s
Now I test an other CPU in this machine, cause on other test machines
with other CPU (AMD Ryzen or others) I get an higher rate.
Unfortunaly the rate on the current machine doesn't rise with the other CPU.
Now I confused if there is any chance to get the restore rate better.
Bye
Gregor
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
2021-10-01 6:52 [PVE-User] proxmox-restore - performance issues Gregor Burck
@ 2021-10-01 7:00 ` Dominik Csapak
0 siblings, 0 replies; 8+ messages in thread
From: Dominik Csapak @ 2021-10-01 7:00 UTC (permalink / raw)
To: Proxmox VE user list, Gregor Burck
On 10/1/21 08:52, Gregor Burck wrote:
> Hi,
hi,
>
> thank you for reply. I made a lot of different tests and setups, but
> this the setup I want to use:
>
> Original setup:
>
> HP DL380 Gen9 with
>
> E5-2640 v3 @ 2.60GHz
> 256 GB RAM
>
> 2x SSDs for host OS
>
> For an ZFS Rais 10:
>
> 2x 1TB SAMSUNG NVME PM983 for spezial devices
> 12x 8 TB HP SAS HDDs
i guess thats the server?
what about the restore client? encryption/sha/etc. will be done by the
client
>
> root@ph-pbs:~# zpool status
> pool: ZFSPOOL
> state: ONLINE
> config:
>
> NAME STATE READ WRITE CKSUM
> ZFSPOOL ONLINE 0 0 0
> mirror-0 ONLINE 0 0 0
> sdc ONLINE 0 0 0
> sdd ONLINE 0 0 0
> mirror-1 ONLINE 0 0 0
> sde ONLINE 0 0 0
> sdf ONLINE 0 0 0
> mirror-2 ONLINE 0 0 0
> sdg ONLINE 0 0 0
> sdh ONLINE 0 0 0
> mirror-3 ONLINE 0 0 0
> sdi ONLINE 0 0 0
> sdj ONLINE 0 0 0
> mirror-4 ONLINE 0 0 0
> sdk ONLINE 0 0 0
> sdl ONLINE 0 0 0
> mirror-5 ONLINE 0 0 0
> sdm ONLINE 0 0 0
> sdn ONLINE 0 0 0
> special
> mirror-6 ONLINE 0 0 0
> nvme0n1 ONLINE 0 0 0
> nvme1n1 ONLINE 0 0 0
>
> errors: No known data errors
>
> pool: rpool
> state: ONLINE
> scan: scrub repaired 0B in 00:02:40 with 0 errors on Sun Aug 8
> 00:26:43 2021
> config:
>
> NAME STATE READ WRITE CKSUM
> rpool ONLINE 0 0 0
> mirror-0 ONLINE 0 0 0
> sda3 ONLINE 0 0 0
> sdb3 ONLINE 0 0 0
>
> errors: No known data errors
>
> The VMSTORE and the BACKUPSTORE is on the zsf as an dataset:
>
> root@ph-pbs:~# zfs list
> NAME USED AVAIL REFER MOUNTPOINT
> ZFSPOOL 10.1T 32.1T 96K /ZFSPOOL
> ZFSPOOL/BACKUPSTORE001 5.63T 32.1T 5.63T /ZFSPOOL/BACKUPSTORE001
> ZFSPOOL/VMSTORE001 4.52T 32.1T 4.52T /ZFSPOOL/VMSTORE001
> rpool 27.3G 80.2G 96K /rpool
> rpool/ROOT 27.3G 80.2G 96K /rpool/ROOT
> rpool/ROOT/pbs-1 27.3G 80.2G 27.3G /
>
> The VM I tested with is our Exchange Server. Raw image size 500GB,
> netto ~400GB content
>
> First Test with one restore job:
>
> Virtual
> Environment 7.0-11
> Datacenter
> Search:
> Logs
> new
> volume ID is 'VMSTORE:vm-101-disk-0'
> restore
> proxmox backup image: /usr/bin/pbs-restore --repository
> root@pam@ph-pbs.peiker-holding.de:ZFSPOOLBACKUP
> vm/121/2021-07-23T19:00:03Z drive-virtio0.img.fidx
> /dev/zvol/ZFSPOOLVMSTORE/vm-101-disk-0 --verbose --format raw
> --skip-zero
> connecting
> to repository 'root@pam@ph-pbs.peiker-holding.de:ZFSPOOLBACKUP'
> open
> block backend for target '/dev/zvol/ZFSPOOLVMSTORE/vm-101-disk-0'
> starting
> to restore snapshot 'vm/121/2021-07-23T19:00:03Z'
> download
> and verify backup index
> progress
> 1% (read 5368709120 bytes, zeroes = 2% (125829120 bytes), duration 86
> sec)
> progress
> 2% (read 10737418240 bytes, zeroes = 1% (159383552 bytes), duration
> 181 sec)
> progress
> 3% (read 16106127360 bytes, zeroes = 0% (159383552 bytes), duration
> 270 sec)
> .
> .
> progress
> 98% (read 526133493760 bytes, zeroes = 0% (3628072960 bytes),
> duration 9492 sec)
> progress
> 99% (read 531502202880 bytes, zeroes = 0% (3628072960 bytes),
> duration 9583 sec)
> progress
> 100% (read 536870912000 bytes, zeroes = 0% (3628072960 bytes),
> duration 9676 sec)
> restore
> image complete (bytes=536870912000, duration=9676.97s,
> speed=52.91MB/s)
> rescan
> volumes...
> TASK
> OK
>
> When I regard iotop I see about the same rate.
>
> But when I start multiple restore jobs parallel, I see that the single
> jon is still on IO 40-50 MB/s but the total IO is multiple of the
> rate. I see on iotop rates to 200-250 MB/s
> So I guess it isn't the store. In some Test with an Setup where I used
> the nvmes as source and target I could reach a singele restore rate
> about 70 MB/s
some disks/storages do not scale with single threaded workloads
(and AFAIR, the pbs-restore must restore a disk single threaded because
of qemu limitations?), but will scale with multiple threads just fine
a 'fio' benchmark of the source as well as the target storage
would be good to get a baseline storage perfomance
see for example: https://pve.proxmox.com/wiki/Benchmarking_Storage
>
> Now I test an other CPU in this machine, cause on other test machines
> with other CPU (AMD Ryzen or others) I get an higher rate.
> Unfortunaly the rate on the current machine doesn't rise with the other
> CPU
can you do a proxmox-backup-client benchmark on all machines and their
respective restore speed (especially the clients; also specify a
repository to see tls speed)
kind regards
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
2021-10-01 9:00 ` Gregor Burck
@ 2021-10-01 9:29 ` Dominik Csapak
0 siblings, 0 replies; 8+ messages in thread
From: Dominik Csapak @ 2021-10-01 9:29 UTC (permalink / raw)
To: Proxmox VE user list, Gregor Burck
On 10/1/21 11:00, Gregor Burck wrote:
> And more information:
>
> I test an fio job, I got the settings not by my own instaed someone from
> the forum give me this for testing ZFS:
is that the source as well as the target storage?
if not please benchmark both
>
>
> root@ph-pbs:~# fio --name=typical-vm --size=8G --rw=readwrite
> --rwmixread=69 --direct=1 --bs=4K --numjobs=4 --ioengine=libaio
> --iodepth=12 --group_reporting --runtime=20m --time_based^C
> root@ph-pbs:~# cd /ZFSPOOL/
> BACKUPSTORE001/ VMSTORE001/
> root@ph-pbs:~# cd /ZFSPOOL/VMSTORE001/
> root@ph-pbs:/ZFSPOOL/VMSTORE001# fio --name=typical-vm --size=8G
> --rw=readwrite --rwmixread=69 --direct=1 --bs=4K --numjobs=4
> --ioengine=libaio --iodepth=12 --group_reporting --runtime=20m --time_based
> typical-vm: (g=0): rw=rw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T)
> 4096B-4096B, ioengine=libaio, iodepth=12
> ...
> fio-3.25
> Starting 4 processes
> typical-vm: Laying out IO file (1 file / 8192MiB)
> typical-vm: Laying out IO file (1 file / 8192MiB)
> typical-vm: Laying out IO file (1 file / 8192MiB)
> typical-vm: Laying out IO file (1 file / 8192MiB)
> Jobs: 4 (f=0): [f(4)][100.0%][r=1518MiB/s,w=682MiB/s][r=389k,w=175k
> IOPS][eta 00m:00s]
> typical-vm: (groupid=0, jobs=4): err= 0: pid=3804786: Fri Oct 1
> 10:56:30 2021
> read: IOPS=356k, BW=1392MiB/s (1460MB/s)(1631GiB/1200001msec)
this looks too high for the storage array, so i guess something is
off with the benchmark (may be cache or missing filename parameter)
and the size is too little (i'd use something that cannot fit
into the cache)
in any case, i'd do read and write benchmarks seperately
as well as setting iodepth and numjobs to 1, to get a baseline
single thread performance
as i wrote in my previous message, check out examples at:
https://pve.proxmox.com/wiki/Benchmarking_Storage
> slat (nsec): min=1854, max=176589k, avg=5156.08, stdev=39010.68
> clat (usec): min=4, max=191637, avg=85.89, stdev=133.21
> lat (usec): min=32, max=191640, avg=91.13, stdev=139.42
> clat percentiles (usec):
> | 1.00th=[ 42], 5.00th=[ 46], 10.00th=[ 49], 20.00th=[ 52],
> | 30.00th=[ 56], 40.00th=[ 59], 50.00th=[ 65], 60.00th=[ 85],
> | 70.00th=[ 97], 80.00th=[ 111], 90.00th=[ 141], 95.00th=[ 176],
> | 99.00th=[ 265], 99.50th=[ 318], 99.90th=[ 570], 99.95th=[ 693],
> | 99.99th=[ 1090]
> bw ( MiB/s): min= 250, max= 2159, per=100.00%, avg=1392.77,
> stdev=63.78, samples=9596
> iops : min=64218, max=552858, avg=356548.75, stdev=16328.20,
> samples=9596
> write: IOPS=160k, BW=626MiB/s (656MB/s)(733GiB/1200001msec); 0 zone
> resets
> slat (usec): min=3, max=191425, avg= 9.71, stdev=34.41
> clat (usec): min=2, max=191641, avg=86.02, stdev=137.32
> lat (usec): min=35, max=191650, avg=95.85, stdev=144.10
> clat percentiles (usec):
> | 1.00th=[ 42], 5.00th=[ 46], 10.00th=[ 49], 20.00th=[ 52],
> | 30.00th=[ 56], 40.00th=[ 59], 50.00th=[ 65], 60.00th=[ 85],
> | 70.00th=[ 98], 80.00th=[ 111], 90.00th=[ 141], 95.00th=[ 178],
> | 99.00th=[ 265], 99.50th=[ 318], 99.90th=[ 578], 99.95th=[ 701],
> | 99.99th=[ 1106]
> bw ( KiB/s): min=114464, max=995856, per=100.00%, avg=640817.51,
> stdev=29342.79, samples=9596
> iops : min=28616, max=248964, avg=160204.26, stdev=7335.70,
> samples=9596
> lat (usec) : 4=0.01%, 10=0.01%, 50=13.69%, 100=58.80%, 250=26.29%
> lat (usec) : 500=1.08%, 750=0.10%, 1000=0.02%
> lat (msec) : 2=0.01%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%
> lat (msec) : 100=0.01%, 250=0.01%
> cpu : usr=18.17%, sys=79.17%, ctx=982498, majf=10, minf=2977
> IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=100.0%, 16=0.0%, 32=0.0%,
> >=64=0.0%
> submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
> >=64=0.0%
> complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.1%, 32=0.0%, 64=0.0%,
> >=64=0.0%
> issued rwts: total=427672030,192161509,0,0 short=0,0,0,0
> dropped=0,0,0,0
> latency : target=0, window=0, percentile=100.00%, depth=12
>
> Run status group 0 (all jobs):
> READ: bw=1392MiB/s (1460MB/s), 1392MiB/s-1392MiB/s
> (1460MB/s-1460MB/s), io=1631GiB (1752GB), run=1200001-1200001msec
> WRITE: bw=626MiB/s (656MB/s), 626MiB/s-626MiB/s (656MB/s-656MB/s),
> io=733GiB (787GB), run=1200001-1200001msec
>
>
> And this is while two of the restore jobs still running.
>
>
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
2021-10-01 7:18 Gregor Burck
@ 2021-10-01 9:00 ` Gregor Burck
2021-10-01 9:29 ` Dominik Csapak
0 siblings, 1 reply; 8+ messages in thread
From: Gregor Burck @ 2021-10-01 9:00 UTC (permalink / raw)
To: Proxmox VE user list
And more information:
I test an fio job, I got the settings not by my own instaed someone
from the forum give me this for testing ZFS:
root@ph-pbs:~# fio --name=typical-vm --size=8G --rw=readwrite
--rwmixread=69 --direct=1 --bs=4K --numjobs=4 --ioengine=libaio
--iodepth=12 --group_reporting --runtime=20m --time_based^C
root@ph-pbs:~# cd /ZFSPOOL/
BACKUPSTORE001/ VMSTORE001/
root@ph-pbs:~# cd /ZFSPOOL/VMSTORE001/
root@ph-pbs:/ZFSPOOL/VMSTORE001# fio --name=typical-vm --size=8G
--rw=readwrite --rwmixread=69 --direct=1 --bs=4K --numjobs=4
--ioengine=libaio --iodepth=12 --group_reporting --runtime=20m
--time_based
typical-vm: (g=0): rw=rw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T)
4096B-4096B, ioengine=libaio, iodepth=12
...
fio-3.25
Starting 4 processes
typical-vm: Laying out IO file (1 file / 8192MiB)
typical-vm: Laying out IO file (1 file / 8192MiB)
typical-vm: Laying out IO file (1 file / 8192MiB)
typical-vm: Laying out IO file (1 file / 8192MiB)
Jobs: 4 (f=0): [f(4)][100.0%][r=1518MiB/s,w=682MiB/s][r=389k,w=175k
IOPS][eta 00m:00s]
typical-vm: (groupid=0, jobs=4): err= 0: pid=3804786: Fri Oct 1 10:56:30 2021
read: IOPS=356k, BW=1392MiB/s (1460MB/s)(1631GiB/1200001msec)
slat (nsec): min=1854, max=176589k, avg=5156.08, stdev=39010.68
clat (usec): min=4, max=191637, avg=85.89, stdev=133.21
lat (usec): min=32, max=191640, avg=91.13, stdev=139.42
clat percentiles (usec):
| 1.00th=[ 42], 5.00th=[ 46], 10.00th=[ 49], 20.00th=[ 52],
| 30.00th=[ 56], 40.00th=[ 59], 50.00th=[ 65], 60.00th=[ 85],
| 70.00th=[ 97], 80.00th=[ 111], 90.00th=[ 141], 95.00th=[ 176],
| 99.00th=[ 265], 99.50th=[ 318], 99.90th=[ 570], 99.95th=[ 693],
| 99.99th=[ 1090]
bw ( MiB/s): min= 250, max= 2159, per=100.00%, avg=1392.77,
stdev=63.78, samples=9596
iops : min=64218, max=552858, avg=356548.75,
stdev=16328.20, samples=9596
write: IOPS=160k, BW=626MiB/s (656MB/s)(733GiB/1200001msec); 0 zone resets
slat (usec): min=3, max=191425, avg= 9.71, stdev=34.41
clat (usec): min=2, max=191641, avg=86.02, stdev=137.32
lat (usec): min=35, max=191650, avg=95.85, stdev=144.10
clat percentiles (usec):
| 1.00th=[ 42], 5.00th=[ 46], 10.00th=[ 49], 20.00th=[ 52],
| 30.00th=[ 56], 40.00th=[ 59], 50.00th=[ 65], 60.00th=[ 85],
| 70.00th=[ 98], 80.00th=[ 111], 90.00th=[ 141], 95.00th=[ 178],
| 99.00th=[ 265], 99.50th=[ 318], 99.90th=[ 578], 99.95th=[ 701],
| 99.99th=[ 1106]
bw ( KiB/s): min=114464, max=995856, per=100.00%, avg=640817.51,
stdev=29342.79, samples=9596
iops : min=28616, max=248964, avg=160204.26, stdev=7335.70,
samples=9596
lat (usec) : 4=0.01%, 10=0.01%, 50=13.69%, 100=58.80%, 250=26.29%
lat (usec) : 500=1.08%, 750=0.10%, 1000=0.02%
lat (msec) : 2=0.01%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%
lat (msec) : 100=0.01%, 250=0.01%
cpu : usr=18.17%, sys=79.17%, ctx=982498, majf=10, minf=2977
IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=100.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.1%, 32=0.0%, 64=0.0%,
>=64=0.0%
issued rwts: total=427672030,192161509,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=12
Run status group 0 (all jobs):
READ: bw=1392MiB/s (1460MB/s), 1392MiB/s-1392MiB/s
(1460MB/s-1460MB/s), io=1631GiB (1752GB), run=1200001-1200001msec
WRITE: bw=626MiB/s (656MB/s), 626MiB/s-626MiB/s (656MB/s-656MB/s),
io=733GiB (787GB), run=1200001-1200001msec
And this is while two of the restore jobs still running.
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
@ 2021-10-01 7:18 Gregor Burck
2021-10-01 9:00 ` Gregor Burck
0 siblings, 1 reply; 8+ messages in thread
From: Gregor Burck @ 2021-10-01 7:18 UTC (permalink / raw)
To: Proxmox VE user list
Hi,
it is PVE and PBS on the same machine, I wrote it on my first mail ;-)
an additional information:
I made in the moment a new test aftar an reboot to exclude to much RAM
iisues on the zfs cache.
With 6 jobs the total io I see over iotop alter between 120MB/s and
240MB/s with peaks of 300 MB/s
When I start htop I see work on all cores but no core is on 100%
and the benchmark:
root@ph-pbs:~# proxmox-backup-client benchmark --repository
root@pam@localhost:BACKUPSTORE001
Password for "root@pam": ************
Uploaded 373 chunks in 5 seconds.
Time per request: 13456 microseconds.
TLS speed: 311.70 MB/s
SHA256 speed: 418.54 MB/s
Compression speed: 572.10 MB/s
Decompress speed: 818.43 MB/s
AES256/GCM speed: 1838.87 MB/s
Verify speed: 298.19 MB/s
┌───────────────────────────────────┬────────────────────┐
│ Name │ Value │
╞═══════════════════════════════════╪════════════════════╡
│ TLS (maximal backup upload speed) │ 311.70 MB/s (25%) │
├───────────────────────────────────┼────────────────────┤
│ SHA256 checksum computation speed │ 418.54 MB/s (21%) │
├───────────────────────────────────┼────────────────────┤
│ ZStd level 1 compression speed │ 572.10 MB/s (76%) │
├───────────────────────────────────┼────────────────────┤
│ ZStd level 1 decompression speed │ 818.43 MB/s (68%) │
├───────────────────────────────────┼────────────────────┤
│ Chunk verification speed │ 298.19 MB/s (39%) │
├───────────────────────────────────┼────────────────────┤
│ AES256 GCM encryption speed │ 1838.87 MB/s (50%) │
└───────────────────────────────────┴────────────────────┘
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
2021-09-30 13:07 ` Gregor Burck
@ 2021-09-30 13:24 ` Dominik Csapak
0 siblings, 0 replies; 8+ messages in thread
From: Dominik Csapak @ 2021-09-30 13:24 UTC (permalink / raw)
To: Proxmox VE user list, Gregor Burck
On 9/30/21 15:07, Gregor Burck wrote:
> Hi,
>
> I made some other test with the same machine but an other proccessor.
>
> I use an Intel(R) Xeon(R) CPU E5-2643 v3 @ 3.40GHz, wich has a higher
> frequency.
>
> The restore rate for an single job dind't change.
>
> Any idea what it could be?
>
> Bye
>
> Gregor
>
>
hi,
can you tell us a bit more about the setup and test?
is the target storage able to handle more than 50MB/s?
how do you measure the 50MB/s?
with kind regards
Dominik
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PVE-User] proxmox-restore - performance issues
2021-09-17 9:29 Gregor Burck
@ 2021-09-30 13:07 ` Gregor Burck
2021-09-30 13:24 ` Dominik Csapak
0 siblings, 1 reply; 8+ messages in thread
From: Gregor Burck @ 2021-09-30 13:07 UTC (permalink / raw)
To: Proxmox VE user list
Hi,
I made some other test with the same machine but an other proccessor.
I use an Intel(R) Xeon(R) CPU E5-2643 v3 @ 3.40GHz, wich has a higher
frequency.
The restore rate for an single job dind't change.
Any idea what it could be?
Bye
Gregor
^ permalink raw reply [flat|nested] 8+ messages in thread
* [PVE-User] proxmox-restore - performance issues
@ 2021-09-17 9:29 Gregor Burck
2021-09-30 13:07 ` Gregor Burck
0 siblings, 1 reply; 8+ messages in thread
From: Gregor Burck @ 2021-09-17 9:29 UTC (permalink / raw)
To: pve-user
Hi,
I've setup an pve/pbs on the same machine:
HP DL380 Gen9
E5-2640 v3 @ 2.60GHz (2 x 8 core)
256 GB RAM
2x 1TB SAMSUNG NVME PM983
12x 8 TB HP SAS HDDs
I create with HDDs and NVME an zfs Raid10.
I still got restore rates of 50 MB/s on one restore job.
If I start multiple jobs parallel the single rate is still on this
rate, but I see with iotop that the summary rate is even higher (max
around 200 MB/s.
When I use htop for the CPU utilisation it seems that an single Job
run only on one core, even when there are multiple tasks.
So I searching the bottle neck, it realy seems not the HDDs.
Any idea so long?
Thank for every,..
Bye
Gregor
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2021-10-01 9:30 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-10-01 6:52 [PVE-User] proxmox-restore - performance issues Gregor Burck
2021-10-01 7:00 ` Dominik Csapak
-- strict thread matches above, loose matches on Subject: below --
2021-10-01 7:18 Gregor Burck
2021-10-01 9:00 ` Gregor Burck
2021-10-01 9:29 ` Dominik Csapak
2021-09-17 9:29 Gregor Burck
2021-09-30 13:07 ` Gregor Burck
2021-09-30 13:24 ` Dominik Csapak
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox