public inbox for pve-devel@lists.proxmox.com
 help / color / mirror / Atom feed
* [pve-devel] [PATCH qemu-server] fix #3588: helper: consider NIC count for config-specific timeout
@ 2024-11-15 10:00 Hannes Laimer
  0 siblings, 0 replies; only message in thread
From: Hannes Laimer @ 2024-11-15 10:00 UTC (permalink / raw)
  To: pve-devel

There have been some reports about `qm start` timeouts on VMs that have a
lot of NICs assigned.
This patch considers the number of NICs when calculating the config-specific
timeout. Since the increase in start time is linearly related to the number
of NICs, a constant timeout increment per NIC was chosen.

Signed-off-by: Hannes Laimer <h.laimer@proxmox.com>
---
 PVE/QemuServer/Helpers.pm | 8 ++++++++
 1 file changed, 8 insertions(+)

diff --git a/PVE/QemuServer/Helpers.pm b/PVE/QemuServer/Helpers.pm
index 0afb6317..99cb3ab8 100644
--- a/PVE/QemuServer/Helpers.pm
+++ b/PVE/QemuServer/Helpers.pm
@@ -167,6 +167,14 @@ sub config_aware_timeout {
 	$timeout = 150;
     }
 
+    # Some testing showed that adding a NIC increased the start time by ~450ms
+    # consistantly across different NIC models, options and already existsing
+    # number of NICs.
+    # So, 10x that to account for any potential system differences seemed
+    # reasonable. 
+    my $nic_count = grep { /^net/ } keys %{$config};
+    $timeout += $nic_count * 5;
+
     return $timeout;
 }
 
-- 
2.39.5



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel


^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2024-11-15 10:00 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-11-15 10:00 [pve-devel] [PATCH qemu-server] fix #3588: helper: consider NIC count for config-specific timeout Hannes Laimer

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal