public inbox for pve-devel@lists.proxmox.com
 help / color / mirror / Atom feed
* [pve-devel] [PATCH ha-manager 0/3] fix mixed resource affinity precedence
@ 2025-09-19 14:08 Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 1/3] tests: add regression tests for mixed resource affinity rules Daniel Kral
                   ` (2 more replies)
  0 siblings, 3 replies; 4+ messages in thread
From: Daniel Kral @ 2025-09-19 14:08 UTC (permalink / raw)
  To: pve-devel


Some last-minute changes to the positive resource affinity's heuristic
(which is described in the next paragraph) in the initial implementation 
broke the mixed usage of resource affinity rules.

Strict positive resource affinity rules narrow down the possible nodes
to a single candidate for a HA resource A, which is the node, where the
most of the HA resources in the positive affinity rule are already
running on and in case of a tie the alphabetically first node is chosen.

If the chosen node contains a HA resource B, which is in negative
affinity with the HA resource A, then there will be no possible node
left as it is removed now as well and won't resolve the incorrect node
placement (see the test cases for more details).


Overall, the guarantees of strict negative resource affinity rules (do
not put negative affinite resources on either the same current nor
migration target node) can become quite expensive/inefficient in the
case where two HA resources, which are in negative resource affinity,
are on the same node initially, because both need to be migrated away
from their current node. Even more so when there aren't enough nodes to
use as "spare" nodes (see the last patch for an example).

This could be resolved in a future patch series by either introducing
non-strict resource affinity rules or loosening that guarantee in the
case where only one HA resource is migrated away and the other stays
put.


Successfully ran `git rebase master --exec 'make clean && make deb'` on
the series before sending.



Daniel Kral (3):
  tests: add regression tests for mixed resource affinity rules
  manager: fix precedence in mixed resource affinity rules usage
  test: add additional mixed resource affinity rule test cases

 src/PVE/HA/Manager.pm                         |  2 +-
 .../README                                    | 16 ++++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  5 +
 .../log.expect                                | 50 ++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              |  7 ++
 .../service_config                            |  6 ++
 .../README                                    | 21 +++++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  5 +
 .../log.expect                                | 68 ++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 +++
 .../service_config                            |  8 ++
 .../README                                    | 20 ++++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  5 +
 .../log.expect                                | 92 +++++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 +++
 .../service_config                            |  8 ++
 .../README                                    | 14 +++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  6 ++
 .../log.expect                                | 85 +++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 +++
 .../service_config                            |  8 ++
 29 files changed, 474 insertions(+), 1 deletion(-)
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/service_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/service_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/service_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/service_config

-- 
2.47.3



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [pve-devel] [PATCH ha-manager 1/3] tests: add regression tests for mixed resource affinity rules
  2025-09-19 14:08 [pve-devel] [PATCH ha-manager 0/3] fix mixed resource affinity precedence Daniel Kral
@ 2025-09-19 14:08 ` Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 2/3] manager: fix precedence in mixed resource affinity rules usage Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 3/3] test: add additional mixed resource affinity rule test cases Daniel Kral
  2 siblings, 0 replies; 4+ messages in thread
From: Daniel Kral @ 2025-09-19 14:08 UTC (permalink / raw)
  To: pve-devel

These test cases show the current behavior of mixed resource affinity
rules, which in the case of test-resource-affinity-strict-mixed2 is
wrong.

Signed-off-by: Daniel Kral <d.kral@proxmox.com>
---
 .../README                                    | 16 ++++++
 .../cmdlist                                   |  3 ++
 .../hardware_status                           |  5 ++
 .../log.expect                                | 50 +++++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              |  7 +++
 .../service_config                            |  6 +++
 .../README                                    | 10 ++++
 .../cmdlist                                   |  3 ++
 .../hardware_status                           |  5 ++
 .../log.expect                                | 48 ++++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 ++++
 .../service_config                            |  8 +++
 14 files changed, 174 insertions(+)
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed1/service_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed2/service_config

diff --git a/src/test/test-resource-affinity-strict-mixed1/README b/src/test/test-resource-affinity-strict-mixed1/README
new file mode 100644
index 00000000..b7003360
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/README
@@ -0,0 +1,16 @@
+The test scenario is:
+- vm:201, vm:202, and vm:203 must be kept together
+- vm:101 and vm:201 must be kept separate
+- Therefore, vm:201, vm:202, vm:203 must all be kept separate from vm:101
+- vm:101 and vm:202 are currently running on node2
+- vm:201 and vm:203 are currently running on node1
+
+The expected outcome is:
+- The resource-node placements do not adhere to the defined resource affinity
+  rules, therefore the HA resources must be moved accordingly: As vm:101 and
+  vm:202 must be on separate nodes, these must be migrated to separate nodes:
+    - As the negative resource affinity rule is strict, resources should
+      neither share the current nor the migration target node, so vm:101 is
+      moved to node2, where neither vm:201, vm:202,nor vm:203 is assigned to
+    - Afterwards, vm:202 is migrated to node1, where vm:201 and vm:203 are
+      already running on
diff --git a/src/test/test-resource-affinity-strict-mixed1/cmdlist b/src/test/test-resource-affinity-strict-mixed1/cmdlist
new file mode 100644
index 00000000..13f90cd7
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/cmdlist
@@ -0,0 +1,3 @@
+[
+    [ "power node1 on", "power node2 on", "power node3 on" ]
+]
diff --git a/src/test/test-resource-affinity-strict-mixed1/hardware_status b/src/test/test-resource-affinity-strict-mixed1/hardware_status
new file mode 100644
index 00000000..451beb13
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/hardware_status
@@ -0,0 +1,5 @@
+{
+  "node1": { "power": "off", "network": "off" },
+  "node2": { "power": "off", "network": "off" },
+  "node3": { "power": "off", "network": "off" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed1/log.expect b/src/test/test-resource-affinity-strict-mixed1/log.expect
new file mode 100644
index 00000000..86e9439f
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/log.expect
@@ -0,0 +1,50 @@
+info      0     hardware: starting simulation
+info     20      cmdlist: execute power node1 on
+info     20    node1/crm: status change startup => wait_for_quorum
+info     20    node1/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node2 on
+info     20    node2/crm: status change startup => wait_for_quorum
+info     20    node2/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node3 on
+info     20    node3/crm: status change startup => wait_for_quorum
+info     20    node3/lrm: status change startup => wait_for_agent_lock
+info     20    node1/crm: got lock 'ha_manager_lock'
+info     20    node1/crm: status change wait_for_quorum => master
+info     20    node1/crm: node 'node1': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node2': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node3': state changed from 'unknown' => 'online'
+info     20    node1/crm: adding new service 'vm:101' on node 'node3'
+info     20    node1/crm: adding new service 'vm:201' on node 'node1'
+info     20    node1/crm: adding new service 'vm:202' on node 'node3'
+info     20    node1/crm: adding new service 'vm:203' on node 'node1'
+info     20    node1/crm: service 'vm:101': state changed from 'request_start' to 'started'  (node = node3)
+info     20    node1/crm: service 'vm:201': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:202': state changed from 'request_start' to 'started'  (node = node3)
+info     20    node1/crm: service 'vm:203': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: migrate service 'vm:101' to node 'node2' (running)
+info     20    node1/crm: service 'vm:101': state changed from 'started' to 'migrate'  (node = node3, target = node2)
+info     20    node1/crm: migrate service 'vm:202' to node 'node1' (running)
+info     20    node1/crm: service 'vm:202': state changed from 'started' to 'migrate'  (node = node3, target = node1)
+info     21    node1/lrm: got lock 'ha_agent_node1_lock'
+info     21    node1/lrm: status change wait_for_agent_lock => active
+info     21    node1/lrm: starting service vm:201
+info     21    node1/lrm: service status vm:201 started
+info     21    node1/lrm: starting service vm:203
+info     21    node1/lrm: service status vm:203 started
+info     22    node2/crm: status change wait_for_quorum => slave
+info     23    node2/lrm: got lock 'ha_agent_node2_lock'
+info     23    node2/lrm: status change wait_for_agent_lock => active
+info     24    node3/crm: status change wait_for_quorum => slave
+info     25    node3/lrm: got lock 'ha_agent_node3_lock'
+info     25    node3/lrm: status change wait_for_agent_lock => active
+info     25    node3/lrm: service vm:101 - start migrate to node 'node2'
+info     25    node3/lrm: service vm:101 - end migrate to node 'node2'
+info     25    node3/lrm: service vm:202 - start migrate to node 'node1'
+info     25    node3/lrm: service vm:202 - end migrate to node 'node1'
+info     40    node1/crm: service 'vm:101': state changed from 'migrate' to 'started'  (node = node2)
+info     40    node1/crm: service 'vm:202': state changed from 'migrate' to 'started'  (node = node1)
+info     41    node1/lrm: starting service vm:202
+info     41    node1/lrm: service status vm:202 started
+info     43    node2/lrm: starting service vm:101
+info     43    node2/lrm: service status vm:101 started
+info    620     hardware: exit simulation - done
diff --git a/src/test/test-resource-affinity-strict-mixed1/manager_status b/src/test/test-resource-affinity-strict-mixed1/manager_status
new file mode 100644
index 00000000..0967ef42
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/manager_status
@@ -0,0 +1 @@
+{}
diff --git a/src/test/test-resource-affinity-strict-mixed1/rules_config b/src/test/test-resource-affinity-strict-mixed1/rules_config
new file mode 100644
index 00000000..2cd9fe21
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/rules_config
@@ -0,0 +1,7 @@
+resource-affinity: vms-100s-must-stick-together
+	resources vm:201,vm:202,vm:203
+	affinity positive
+
+resource-affinity: vms-100s-and-vm201-must-be-separate
+	resources vm:201,vm:101
+	affinity negative
diff --git a/src/test/test-resource-affinity-strict-mixed1/service_config b/src/test/test-resource-affinity-strict-mixed1/service_config
new file mode 100644
index 00000000..83e2157d
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed1/service_config
@@ -0,0 +1,6 @@
+{
+    "vm:101": { "node": "node3", "state": "started" },
+    "vm:201": { "node": "node1", "state": "started" },
+    "vm:202": { "node": "node3", "state": "started" },
+    "vm:203": { "node": "node1", "state": "started" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed2/README b/src/test/test-resource-affinity-strict-mixed2/README
new file mode 100644
index 00000000..c56d1a2d
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/README
@@ -0,0 +1,10 @@
+The test scenario is:
+- vm:101, vm:102, and vm:103 must be kept together
+- vm:201, vm:202, and vm:203 must be kept together
+- vm:101 and vm:201 must be kept separate
+- Therefore, vm:101, vm:102, and vm:103 must all be kept separate from vm:201,
+  vm:202, and vm:203 and vice versa
+- vm:101, vm:103, vm:201, and vm:203 are currently running on node1
+- vm:102 and vm:202 are running on node3 and node2 respectively
+
+The current outcome is incorrect.
diff --git a/src/test/test-resource-affinity-strict-mixed2/cmdlist b/src/test/test-resource-affinity-strict-mixed2/cmdlist
new file mode 100644
index 00000000..13f90cd7
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/cmdlist
@@ -0,0 +1,3 @@
+[
+    [ "power node1 on", "power node2 on", "power node3 on" ]
+]
diff --git a/src/test/test-resource-affinity-strict-mixed2/hardware_status b/src/test/test-resource-affinity-strict-mixed2/hardware_status
new file mode 100644
index 00000000..451beb13
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/hardware_status
@@ -0,0 +1,5 @@
+{
+  "node1": { "power": "off", "network": "off" },
+  "node2": { "power": "off", "network": "off" },
+  "node3": { "power": "off", "network": "off" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed2/log.expect b/src/test/test-resource-affinity-strict-mixed2/log.expect
new file mode 100644
index 00000000..9cdc8b14
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/log.expect
@@ -0,0 +1,48 @@
+info      0     hardware: starting simulation
+info     20      cmdlist: execute power node1 on
+info     20    node1/crm: status change startup => wait_for_quorum
+info     20    node1/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node2 on
+info     20    node2/crm: status change startup => wait_for_quorum
+info     20    node2/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node3 on
+info     20    node3/crm: status change startup => wait_for_quorum
+info     20    node3/lrm: status change startup => wait_for_agent_lock
+info     20    node1/crm: got lock 'ha_manager_lock'
+info     20    node1/crm: status change wait_for_quorum => master
+info     20    node1/crm: node 'node1': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node2': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node3': state changed from 'unknown' => 'online'
+info     20    node1/crm: adding new service 'vm:101' on node 'node1'
+info     20    node1/crm: adding new service 'vm:102' on node 'node3'
+info     20    node1/crm: adding new service 'vm:103' on node 'node1'
+info     20    node1/crm: adding new service 'vm:201' on node 'node1'
+info     20    node1/crm: adding new service 'vm:202' on node 'node2'
+info     20    node1/crm: adding new service 'vm:203' on node 'node1'
+info     20    node1/crm: service 'vm:101': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:102': state changed from 'request_start' to 'started'  (node = node3)
+info     20    node1/crm: service 'vm:103': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:201': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:202': state changed from 'request_start' to 'started'  (node = node2)
+info     20    node1/crm: service 'vm:203': state changed from 'request_start' to 'started'  (node = node1)
+info     21    node1/lrm: got lock 'ha_agent_node1_lock'
+info     21    node1/lrm: status change wait_for_agent_lock => active
+info     21    node1/lrm: starting service vm:101
+info     21    node1/lrm: service status vm:101 started
+info     21    node1/lrm: starting service vm:103
+info     21    node1/lrm: service status vm:103 started
+info     21    node1/lrm: starting service vm:201
+info     21    node1/lrm: service status vm:201 started
+info     21    node1/lrm: starting service vm:203
+info     21    node1/lrm: service status vm:203 started
+info     22    node2/crm: status change wait_for_quorum => slave
+info     23    node2/lrm: got lock 'ha_agent_node2_lock'
+info     23    node2/lrm: status change wait_for_agent_lock => active
+info     23    node2/lrm: starting service vm:202
+info     23    node2/lrm: service status vm:202 started
+info     24    node3/crm: status change wait_for_quorum => slave
+info     25    node3/lrm: got lock 'ha_agent_node3_lock'
+info     25    node3/lrm: status change wait_for_agent_lock => active
+info     25    node3/lrm: starting service vm:102
+info     25    node3/lrm: service status vm:102 started
+info    620     hardware: exit simulation - done
diff --git a/src/test/test-resource-affinity-strict-mixed2/manager_status b/src/test/test-resource-affinity-strict-mixed2/manager_status
new file mode 100644
index 00000000..0967ef42
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/manager_status
@@ -0,0 +1 @@
+{}
diff --git a/src/test/test-resource-affinity-strict-mixed2/rules_config b/src/test/test-resource-affinity-strict-mixed2/rules_config
new file mode 100644
index 00000000..851ed590
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/rules_config
@@ -0,0 +1,11 @@
+resource-affinity: together-100s
+	resources vm:101,vm:102,vm:103
+	affinity positive
+
+resource-affinity: together-200s
+	resources vm:201,vm:202,vm:203
+	affinity positive
+
+resource-affinity: lonely-must-vms-be
+	resources vm:101,vm:201
+	affinity negative
diff --git a/src/test/test-resource-affinity-strict-mixed2/service_config b/src/test/test-resource-affinity-strict-mixed2/service_config
new file mode 100644
index 00000000..fe6b2438
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed2/service_config
@@ -0,0 +1,8 @@
+{
+    "vm:101": { "node": "node1", "state": "started" },
+    "vm:102": { "node": "node3", "state": "started" },
+    "vm:103": { "node": "node1", "state": "started" },
+    "vm:201": { "node": "node1", "state": "started" },
+    "vm:202": { "node": "node2", "state": "started" },
+    "vm:203": { "node": "node1", "state": "started" }
+}
-- 
2.47.3



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [pve-devel] [PATCH ha-manager 2/3] manager: fix precedence in mixed resource affinity rules usage
  2025-09-19 14:08 [pve-devel] [PATCH ha-manager 0/3] fix mixed resource affinity precedence Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 1/3] tests: add regression tests for mixed resource affinity rules Daniel Kral
@ 2025-09-19 14:08 ` Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 3/3] test: add additional mixed resource affinity rule test cases Daniel Kral
  2 siblings, 0 replies; 4+ messages in thread
From: Daniel Kral @ 2025-09-19 14:08 UTC (permalink / raw)
  To: pve-devel

Strict positive resource affinity rules narrow down the possible nodes
to a single candidate for a HA resource A, which is the node, where the
most of the HA resources in the positive affinity rule are already
running on and in case of a tie the alphabetically first node is chosen.

If the chosen node contains a HA resource B, which is in negative
affinity with the HA resource A, then $pri_nodes will become empty and
will not result in any migration.

Therefore, apply the negative resource affinity rules before the
positive resource affinity rules to prevent the premature pruning of
nodes.

Signed-off-by: Daniel Kral <d.kral@proxmox.com>
---
 src/PVE/HA/Manager.pm                         |  2 +-
 .../README                                    | 13 ++++++-
 .../log.expect                                | 36 ++++++++++++++-----
 3 files changed, 41 insertions(+), 10 deletions(-)

diff --git a/src/PVE/HA/Manager.pm b/src/PVE/HA/Manager.pm
index ba59f642..3d74288d 100644
--- a/src/PVE/HA/Manager.pm
+++ b/src/PVE/HA/Manager.pm
@@ -176,8 +176,8 @@ sub select_service_node {
         }
     }
 
-    apply_positive_resource_affinity($together, $pri_nodes);
     apply_negative_resource_affinity($separate, $pri_nodes);
+    apply_positive_resource_affinity($together, $pri_nodes);
 
     return $maintenance_fallback
         if defined($maintenance_fallback) && $pri_nodes->{$maintenance_fallback};
diff --git a/src/test/test-resource-affinity-strict-mixed2/README b/src/test/test-resource-affinity-strict-mixed2/README
index c56d1a2d..a4d89ff3 100644
--- a/src/test/test-resource-affinity-strict-mixed2/README
+++ b/src/test/test-resource-affinity-strict-mixed2/README
@@ -7,4 +7,15 @@ The test scenario is:
 - vm:101, vm:103, vm:201, and vm:203 are currently running on node1
 - vm:102 and vm:202 are running on node3 and node2 respectively
 
-The current outcome is incorrect.
+The expected outcome is:
+- The resource-node placements do not adhere to the defined resource affinity
+  rules, therefore the HA resources must be moved accordingly: As vm:101 and
+  vm:103 must be kept separate from vm:201 and vm:203, which are all currently
+  running on node1, these must be migrated to separate nodes:
+    - As the negative resource affinity rule is strict, resources must neither
+      share the current nor the migration target node, so both positive
+      affinity groups must be put on "spare" nodes, which in that case is node3
+      (for vm:101 and vm:103) and node2 (for vm:201 and vm:203) respectively.
+      These node selections are because there are already other positive
+      resource affinity rule members running on these nodes (vm:102 on node3
+      and vm:202 on node2).
diff --git a/src/test/test-resource-affinity-strict-mixed2/log.expect b/src/test/test-resource-affinity-strict-mixed2/log.expect
index 9cdc8b14..e7081e4b 100644
--- a/src/test/test-resource-affinity-strict-mixed2/log.expect
+++ b/src/test/test-resource-affinity-strict-mixed2/log.expect
@@ -25,16 +25,24 @@ info     20    node1/crm: service 'vm:103': state changed from 'request_start' t
 info     20    node1/crm: service 'vm:201': state changed from 'request_start' to 'started'  (node = node1)
 info     20    node1/crm: service 'vm:202': state changed from 'request_start' to 'started'  (node = node2)
 info     20    node1/crm: service 'vm:203': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: migrate service 'vm:101' to node 'node3' (running)
+info     20    node1/crm: service 'vm:101': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:103' to node 'node3' (running)
+info     20    node1/crm: service 'vm:103': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:201' to node 'node2' (running)
+info     20    node1/crm: service 'vm:201': state changed from 'started' to 'migrate'  (node = node1, target = node2)
+info     20    node1/crm: migrate service 'vm:203' to node 'node2' (running)
+info     20    node1/crm: service 'vm:203': state changed from 'started' to 'migrate'  (node = node1, target = node2)
 info     21    node1/lrm: got lock 'ha_agent_node1_lock'
 info     21    node1/lrm: status change wait_for_agent_lock => active
-info     21    node1/lrm: starting service vm:101
-info     21    node1/lrm: service status vm:101 started
-info     21    node1/lrm: starting service vm:103
-info     21    node1/lrm: service status vm:103 started
-info     21    node1/lrm: starting service vm:201
-info     21    node1/lrm: service status vm:201 started
-info     21    node1/lrm: starting service vm:203
-info     21    node1/lrm: service status vm:203 started
+info     21    node1/lrm: service vm:101 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:101 - end migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - end migrate to node 'node3'
+info     21    node1/lrm: service vm:201 - start migrate to node 'node2'
+info     21    node1/lrm: service vm:201 - end migrate to node 'node2'
+info     21    node1/lrm: service vm:203 - start migrate to node 'node2'
+info     21    node1/lrm: service vm:203 - end migrate to node 'node2'
 info     22    node2/crm: status change wait_for_quorum => slave
 info     23    node2/lrm: got lock 'ha_agent_node2_lock'
 info     23    node2/lrm: status change wait_for_agent_lock => active
@@ -45,4 +53,16 @@ info     25    node3/lrm: got lock 'ha_agent_node3_lock'
 info     25    node3/lrm: status change wait_for_agent_lock => active
 info     25    node3/lrm: starting service vm:102
 info     25    node3/lrm: service status vm:102 started
+info     40    node1/crm: service 'vm:101': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:103': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:201': state changed from 'migrate' to 'started'  (node = node2)
+info     40    node1/crm: service 'vm:203': state changed from 'migrate' to 'started'  (node = node2)
+info     43    node2/lrm: starting service vm:201
+info     43    node2/lrm: service status vm:201 started
+info     43    node2/lrm: starting service vm:203
+info     43    node2/lrm: service status vm:203 started
+info     45    node3/lrm: starting service vm:101
+info     45    node3/lrm: service status vm:101 started
+info     45    node3/lrm: starting service vm:103
+info     45    node3/lrm: service status vm:103 started
 info    620     hardware: exit simulation - done
-- 
2.47.3



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [pve-devel] [PATCH ha-manager 3/3] test: add additional mixed resource affinity rule test cases
  2025-09-19 14:08 [pve-devel] [PATCH ha-manager 0/3] fix mixed resource affinity precedence Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 1/3] tests: add regression tests for mixed resource affinity rules Daniel Kral
  2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 2/3] manager: fix precedence in mixed resource affinity rules usage Daniel Kral
@ 2025-09-19 14:08 ` Daniel Kral
  2 siblings, 0 replies; 4+ messages in thread
From: Daniel Kral @ 2025-09-19 14:08 UTC (permalink / raw)
  To: pve-devel

These first test case shows and documents changes to the case, where
there aren't enough nodes to properly hold the guarantees of strict
resource affinity rules nor are efficiently solved yet.

The second test case shows that the former test case would have needed
four nodes to hold all guarantees and resolve the wrong node placement
in one step.

Signed-off-by: Daniel Kral <d.kral@proxmox.com>
---
 .../README                                    | 20 ++++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  5 +
 .../log.expect                                | 92 +++++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 +++
 .../service_config                            |  8 ++
 .../README                                    | 14 +++
 .../cmdlist                                   |  3 +
 .../hardware_status                           |  6 ++
 .../log.expect                                | 85 +++++++++++++++++
 .../manager_status                            |  1 +
 .../rules_config                              | 11 +++
 .../service_config                            |  8 ++
 14 files changed, 268 insertions(+)
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed3/service_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/README
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/cmdlist
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/hardware_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/log.expect
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/manager_status
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/rules_config
 create mode 100644 src/test/test-resource-affinity-strict-mixed4/service_config

diff --git a/src/test/test-resource-affinity-strict-mixed3/README b/src/test/test-resource-affinity-strict-mixed3/README
new file mode 100644
index 00000000..dc8ec152
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/README
@@ -0,0 +1,20 @@
+The test scenario is:
+- vm:101, vm:102, and vm:103 must be kept together
+- vm:201, vm:202, and vm:203 must be kept together
+- vm:101 and vm:201 must be kept separate
+- Therefore, vm:101, vm:102, and vm:103 must all be kept separate from vm:201,
+  vm:202, and vm:203 and vice versa
+- vm:101, vm:103, vm:201, and vm:203 are currently running on node1
+- vm:102 and vm:202 are both running on node2
+
+The expected outcome is:
+- The resource-node placements do not adhere to the defined resource affinity
+  rules, therefore the HA resources must be moved accordingly. In the end,
+  vm:101, vm:102, and vm:103 should be on a separate node as vm:201, vm:202,
+  and vm:203.
+
+The current final outcome is correct, but is inefficient and doesn't hold all
+guarantees (i.e. do not put resources in strict node affinity rules on the same
+node nor their migration target) at all times. As shown by
+test-resources-affinity-strict-mixed4, it needs at least four nodes to hold all
+guarantees and rebalance the resources to their correct node placements.
diff --git a/src/test/test-resource-affinity-strict-mixed3/cmdlist b/src/test/test-resource-affinity-strict-mixed3/cmdlist
new file mode 100644
index 00000000..13f90cd7
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/cmdlist
@@ -0,0 +1,3 @@
+[
+    [ "power node1 on", "power node2 on", "power node3 on" ]
+]
diff --git a/src/test/test-resource-affinity-strict-mixed3/hardware_status b/src/test/test-resource-affinity-strict-mixed3/hardware_status
new file mode 100644
index 00000000..451beb13
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/hardware_status
@@ -0,0 +1,5 @@
+{
+  "node1": { "power": "off", "network": "off" },
+  "node2": { "power": "off", "network": "off" },
+  "node3": { "power": "off", "network": "off" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed3/log.expect b/src/test/test-resource-affinity-strict-mixed3/log.expect
new file mode 100644
index 00000000..b3de104f
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/log.expect
@@ -0,0 +1,92 @@
+info      0     hardware: starting simulation
+info     20      cmdlist: execute power node1 on
+info     20    node1/crm: status change startup => wait_for_quorum
+info     20    node1/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node2 on
+info     20    node2/crm: status change startup => wait_for_quorum
+info     20    node2/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node3 on
+info     20    node3/crm: status change startup => wait_for_quorum
+info     20    node3/lrm: status change startup => wait_for_agent_lock
+info     20    node1/crm: got lock 'ha_manager_lock'
+info     20    node1/crm: status change wait_for_quorum => master
+info     20    node1/crm: node 'node1': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node2': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node3': state changed from 'unknown' => 'online'
+info     20    node1/crm: adding new service 'vm:101' on node 'node1'
+info     20    node1/crm: adding new service 'vm:102' on node 'node2'
+info     20    node1/crm: adding new service 'vm:103' on node 'node1'
+info     20    node1/crm: adding new service 'vm:201' on node 'node1'
+info     20    node1/crm: adding new service 'vm:202' on node 'node2'
+info     20    node1/crm: adding new service 'vm:203' on node 'node1'
+info     20    node1/crm: service 'vm:101': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:102': state changed from 'request_start' to 'started'  (node = node2)
+info     20    node1/crm: service 'vm:103': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:201': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:202': state changed from 'request_start' to 'started'  (node = node2)
+info     20    node1/crm: service 'vm:203': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: migrate service 'vm:101' to node 'node3' (running)
+info     20    node1/crm: service 'vm:101': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:102' to node 'node3' (running)
+info     20    node1/crm: service 'vm:102': state changed from 'started' to 'migrate'  (node = node2, target = node3)
+info     20    node1/crm: migrate service 'vm:103' to node 'node3' (running)
+info     20    node1/crm: service 'vm:103': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:203' to node 'node2' (running)
+info     20    node1/crm: service 'vm:203': state changed from 'started' to 'migrate'  (node = node1, target = node2)
+info     21    node1/lrm: got lock 'ha_agent_node1_lock'
+info     21    node1/lrm: status change wait_for_agent_lock => active
+info     21    node1/lrm: service vm:101 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:101 - end migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - end migrate to node 'node3'
+info     21    node1/lrm: starting service vm:201
+info     21    node1/lrm: service status vm:201 started
+info     21    node1/lrm: service vm:203 - start migrate to node 'node2'
+info     21    node1/lrm: service vm:203 - end migrate to node 'node2'
+info     22    node2/crm: status change wait_for_quorum => slave
+info     23    node2/lrm: got lock 'ha_agent_node2_lock'
+info     23    node2/lrm: status change wait_for_agent_lock => active
+info     23    node2/lrm: service vm:102 - start migrate to node 'node3'
+info     23    node2/lrm: service vm:102 - end migrate to node 'node3'
+info     23    node2/lrm: starting service vm:202
+info     23    node2/lrm: service status vm:202 started
+info     24    node3/crm: status change wait_for_quorum => slave
+info     25    node3/lrm: got lock 'ha_agent_node3_lock'
+info     25    node3/lrm: status change wait_for_agent_lock => active
+info     40    node1/crm: service 'vm:101': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:102': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:103': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: migrate service 'vm:201' to node 'node2' (running)
+info     40    node1/crm: service 'vm:201': state changed from 'started' to 'migrate'  (node = node1, target = node2)
+info     40    node1/crm: migrate service 'vm:202' to node 'node1' (running)
+info     40    node1/crm: service 'vm:202': state changed from 'started' to 'migrate'  (node = node2, target = node1)
+info     40    node1/crm: service 'vm:203': state changed from 'migrate' to 'started'  (node = node2)
+info     40    node1/crm: migrate service 'vm:203' to node 'node1' (running)
+info     40    node1/crm: service 'vm:203': state changed from 'started' to 'migrate'  (node = node2, target = node1)
+info     41    node1/lrm: service vm:201 - start migrate to node 'node2'
+info     41    node1/lrm: service vm:201 - end migrate to node 'node2'
+info     43    node2/lrm: service vm:202 - start migrate to node 'node1'
+info     43    node2/lrm: service vm:202 - end migrate to node 'node1'
+info     43    node2/lrm: service vm:203 - start migrate to node 'node1'
+info     43    node2/lrm: service vm:203 - end migrate to node 'node1'
+info     45    node3/lrm: starting service vm:101
+info     45    node3/lrm: service status vm:101 started
+info     45    node3/lrm: starting service vm:102
+info     45    node3/lrm: service status vm:102 started
+info     45    node3/lrm: starting service vm:103
+info     45    node3/lrm: service status vm:103 started
+info     60    node1/crm: service 'vm:201': state changed from 'migrate' to 'started'  (node = node2)
+info     60    node1/crm: service 'vm:202': state changed from 'migrate' to 'started'  (node = node1)
+info     60    node1/crm: service 'vm:203': state changed from 'migrate' to 'started'  (node = node1)
+info     60    node1/crm: migrate service 'vm:201' to node 'node1' (running)
+info     60    node1/crm: service 'vm:201': state changed from 'started' to 'migrate'  (node = node2, target = node1)
+info     61    node1/lrm: starting service vm:202
+info     61    node1/lrm: service status vm:202 started
+info     61    node1/lrm: starting service vm:203
+info     61    node1/lrm: service status vm:203 started
+info     63    node2/lrm: service vm:201 - start migrate to node 'node1'
+info     63    node2/lrm: service vm:201 - end migrate to node 'node1'
+info     80    node1/crm: service 'vm:201': state changed from 'migrate' to 'started'  (node = node1)
+info     81    node1/lrm: starting service vm:201
+info     81    node1/lrm: service status vm:201 started
+info    620     hardware: exit simulation - done
diff --git a/src/test/test-resource-affinity-strict-mixed3/manager_status b/src/test/test-resource-affinity-strict-mixed3/manager_status
new file mode 100644
index 00000000..0967ef42
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/manager_status
@@ -0,0 +1 @@
+{}
diff --git a/src/test/test-resource-affinity-strict-mixed3/rules_config b/src/test/test-resource-affinity-strict-mixed3/rules_config
new file mode 100644
index 00000000..851ed590
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/rules_config
@@ -0,0 +1,11 @@
+resource-affinity: together-100s
+	resources vm:101,vm:102,vm:103
+	affinity positive
+
+resource-affinity: together-200s
+	resources vm:201,vm:202,vm:203
+	affinity positive
+
+resource-affinity: lonely-must-vms-be
+	resources vm:101,vm:201
+	affinity negative
diff --git a/src/test/test-resource-affinity-strict-mixed3/service_config b/src/test/test-resource-affinity-strict-mixed3/service_config
new file mode 100644
index 00000000..3028810b
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed3/service_config
@@ -0,0 +1,8 @@
+{
+    "vm:101": { "node": "node1", "state": "started" },
+    "vm:102": { "node": "node2", "state": "started" },
+    "vm:103": { "node": "node1", "state": "started" },
+    "vm:201": { "node": "node1", "state": "started" },
+    "vm:202": { "node": "node2", "state": "started" },
+    "vm:203": { "node": "node1", "state": "started" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed4/README b/src/test/test-resource-affinity-strict-mixed4/README
new file mode 100644
index 00000000..25e5abc7
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/README
@@ -0,0 +1,14 @@
+The test scenario is:
+- vm:101, vm:102, and vm:103 must be kept together
+- vm:201, vm:202, and vm:203 must be kept together
+- vm:101 and vm:201 must be kept separate
+- Therefore, vm:101, vm:102, and vm:103 must all be kept separate from vm:201,
+  vm:202, and vm:203 and vice versa
+- vm:101, vm:103, vm:201, and vm:203 are currently running on node1
+- vm:102 and vm:202 are both running on node2
+
+The expected outcome is:
+- The resource-node placements do not adhere to the defined resource affinity
+  rules, therefore the HA resources must be moved accordingly. In the end,
+  vm:101, vm:102, and vm:103 should be on a separate node as vm:201, vm:202,
+  and vm:203.
diff --git a/src/test/test-resource-affinity-strict-mixed4/cmdlist b/src/test/test-resource-affinity-strict-mixed4/cmdlist
new file mode 100644
index 00000000..043a94a6
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/cmdlist
@@ -0,0 +1,3 @@
+[
+    [ "power node1 on", "power node2 on", "power node3 on", "power node4 on" ]
+]
diff --git a/src/test/test-resource-affinity-strict-mixed4/hardware_status b/src/test/test-resource-affinity-strict-mixed4/hardware_status
new file mode 100644
index 00000000..4aed08a1
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/hardware_status
@@ -0,0 +1,6 @@
+{
+  "node1": { "power": "off", "network": "off" },
+  "node2": { "power": "off", "network": "off" },
+  "node3": { "power": "off", "network": "off" },
+  "node4": { "power": "off", "network": "off" }
+}
diff --git a/src/test/test-resource-affinity-strict-mixed4/log.expect b/src/test/test-resource-affinity-strict-mixed4/log.expect
new file mode 100644
index 00000000..903af623
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/log.expect
@@ -0,0 +1,85 @@
+info      0     hardware: starting simulation
+info     20      cmdlist: execute power node1 on
+info     20    node1/crm: status change startup => wait_for_quorum
+info     20    node1/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node2 on
+info     20    node2/crm: status change startup => wait_for_quorum
+info     20    node2/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node3 on
+info     20    node3/crm: status change startup => wait_for_quorum
+info     20    node3/lrm: status change startup => wait_for_agent_lock
+info     20      cmdlist: execute power node4 on
+info     20    node4/crm: status change startup => wait_for_quorum
+info     20    node4/lrm: status change startup => wait_for_agent_lock
+info     20    node1/crm: got lock 'ha_manager_lock'
+info     20    node1/crm: status change wait_for_quorum => master
+info     20    node1/crm: node 'node1': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node2': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node3': state changed from 'unknown' => 'online'
+info     20    node1/crm: node 'node4': state changed from 'unknown' => 'online'
+info     20    node1/crm: adding new service 'vm:101' on node 'node1'
+info     20    node1/crm: adding new service 'vm:102' on node 'node2'
+info     20    node1/crm: adding new service 'vm:103' on node 'node1'
+info     20    node1/crm: adding new service 'vm:201' on node 'node1'
+info     20    node1/crm: adding new service 'vm:202' on node 'node2'
+info     20    node1/crm: adding new service 'vm:203' on node 'node1'
+info     20    node1/crm: service 'vm:101': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:102': state changed from 'request_start' to 'started'  (node = node2)
+info     20    node1/crm: service 'vm:103': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:201': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: service 'vm:202': state changed from 'request_start' to 'started'  (node = node2)
+info     20    node1/crm: service 'vm:203': state changed from 'request_start' to 'started'  (node = node1)
+info     20    node1/crm: migrate service 'vm:101' to node 'node3' (running)
+info     20    node1/crm: service 'vm:101': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:102' to node 'node3' (running)
+info     20    node1/crm: service 'vm:102': state changed from 'started' to 'migrate'  (node = node2, target = node3)
+info     20    node1/crm: migrate service 'vm:103' to node 'node3' (running)
+info     20    node1/crm: service 'vm:103': state changed from 'started' to 'migrate'  (node = node1, target = node3)
+info     20    node1/crm: migrate service 'vm:201' to node 'node4' (running)
+info     20    node1/crm: service 'vm:201': state changed from 'started' to 'migrate'  (node = node1, target = node4)
+info     20    node1/crm: migrate service 'vm:202' to node 'node4' (running)
+info     20    node1/crm: service 'vm:202': state changed from 'started' to 'migrate'  (node = node2, target = node4)
+info     20    node1/crm: migrate service 'vm:203' to node 'node4' (running)
+info     20    node1/crm: service 'vm:203': state changed from 'started' to 'migrate'  (node = node1, target = node4)
+info     21    node1/lrm: got lock 'ha_agent_node1_lock'
+info     21    node1/lrm: status change wait_for_agent_lock => active
+info     21    node1/lrm: service vm:101 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:101 - end migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - start migrate to node 'node3'
+info     21    node1/lrm: service vm:103 - end migrate to node 'node3'
+info     21    node1/lrm: service vm:201 - start migrate to node 'node4'
+info     21    node1/lrm: service vm:201 - end migrate to node 'node4'
+info     21    node1/lrm: service vm:203 - start migrate to node 'node4'
+info     21    node1/lrm: service vm:203 - end migrate to node 'node4'
+info     22    node2/crm: status change wait_for_quorum => slave
+info     23    node2/lrm: got lock 'ha_agent_node2_lock'
+info     23    node2/lrm: status change wait_for_agent_lock => active
+info     23    node2/lrm: service vm:102 - start migrate to node 'node3'
+info     23    node2/lrm: service vm:102 - end migrate to node 'node3'
+info     23    node2/lrm: service vm:202 - start migrate to node 'node4'
+info     23    node2/lrm: service vm:202 - end migrate to node 'node4'
+info     24    node3/crm: status change wait_for_quorum => slave
+info     25    node3/lrm: got lock 'ha_agent_node3_lock'
+info     25    node3/lrm: status change wait_for_agent_lock => active
+info     26    node4/crm: status change wait_for_quorum => slave
+info     27    node4/lrm: got lock 'ha_agent_node4_lock'
+info     27    node4/lrm: status change wait_for_agent_lock => active
+info     40    node1/crm: service 'vm:101': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:102': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:103': state changed from 'migrate' to 'started'  (node = node3)
+info     40    node1/crm: service 'vm:201': state changed from 'migrate' to 'started'  (node = node4)
+info     40    node1/crm: service 'vm:202': state changed from 'migrate' to 'started'  (node = node4)
+info     40    node1/crm: service 'vm:203': state changed from 'migrate' to 'started'  (node = node4)
+info     45    node3/lrm: starting service vm:101
+info     45    node3/lrm: service status vm:101 started
+info     45    node3/lrm: starting service vm:102
+info     45    node3/lrm: service status vm:102 started
+info     45    node3/lrm: starting service vm:103
+info     45    node3/lrm: service status vm:103 started
+info     47    node4/lrm: starting service vm:201
+info     47    node4/lrm: service status vm:201 started
+info     47    node4/lrm: starting service vm:202
+info     47    node4/lrm: service status vm:202 started
+info     47    node4/lrm: starting service vm:203
+info     47    node4/lrm: service status vm:203 started
+info    620     hardware: exit simulation - done
diff --git a/src/test/test-resource-affinity-strict-mixed4/manager_status b/src/test/test-resource-affinity-strict-mixed4/manager_status
new file mode 100644
index 00000000..0967ef42
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/manager_status
@@ -0,0 +1 @@
+{}
diff --git a/src/test/test-resource-affinity-strict-mixed4/rules_config b/src/test/test-resource-affinity-strict-mixed4/rules_config
new file mode 100644
index 00000000..851ed590
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/rules_config
@@ -0,0 +1,11 @@
+resource-affinity: together-100s
+	resources vm:101,vm:102,vm:103
+	affinity positive
+
+resource-affinity: together-200s
+	resources vm:201,vm:202,vm:203
+	affinity positive
+
+resource-affinity: lonely-must-vms-be
+	resources vm:101,vm:201
+	affinity negative
diff --git a/src/test/test-resource-affinity-strict-mixed4/service_config b/src/test/test-resource-affinity-strict-mixed4/service_config
new file mode 100644
index 00000000..3028810b
--- /dev/null
+++ b/src/test/test-resource-affinity-strict-mixed4/service_config
@@ -0,0 +1,8 @@
+{
+    "vm:101": { "node": "node1", "state": "started" },
+    "vm:102": { "node": "node2", "state": "started" },
+    "vm:103": { "node": "node1", "state": "started" },
+    "vm:201": { "node": "node1", "state": "started" },
+    "vm:202": { "node": "node2", "state": "started" },
+    "vm:203": { "node": "node1", "state": "started" }
+}
-- 
2.47.3



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel


^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2025-09-19 14:09 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2025-09-19 14:08 [pve-devel] [PATCH ha-manager 0/3] fix mixed resource affinity precedence Daniel Kral
2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 1/3] tests: add regression tests for mixed resource affinity rules Daniel Kral
2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 2/3] manager: fix precedence in mixed resource affinity rules usage Daniel Kral
2025-09-19 14:08 ` [pve-devel] [PATCH ha-manager 3/3] test: add additional mixed resource affinity rule test cases Daniel Kral

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal