* [PVE-User] VM Migration Not Happening :-(
@ 2024-09-01 8:10 duluxoz
2024-09-01 17:14 ` Lund Svein-Erik
2024-09-01 22:02 ` Gilberto Ferreira
0 siblings, 2 replies; 6+ messages in thread
From: duluxoz @ 2024-09-01 8:10 UTC (permalink / raw)
To: Proxmox VE user list
Hi All,
I need help with figuring out why I can't migrate a VM from one Proxmox
Node to another (in the same cluster, of course).
These are the details provided by the Proxmox Task Log:
```
task started by HA resource agent
2024-09-01 18:02:30 use dedicated network address for sending migration
traffic (192.168.200.103)
2024-09-01 18:02:30 starting migration of VM 100 to node 'pven3'
(192.168.200.103)
2024-09-01 18:02:30 starting VM 100 on remote node 'pven3'
2024-09-01 18:02:30 [pven3]
2024-09-01 18:02:32 start remote tunnel
2024-09-01 18:02:33 ssh tunnel ver 1
2024-09-01 18:02:33 starting online/live migration on
unix:/run/qemu-server/100.migrate
2024-09-01 18:02:33 set migration capabilities
2024-09-01 18:02:33 migration downtime limit: 100 ms
2024-09-01 18:02:33 migration cachesize: 256.0 MiB
2024-09-01 18:02:33 set migration parameters
2024-09-01 18:02:33 start migrate command to
unix:/run/qemu-server/100.migrate
channel 2: open failed: connect failed: open failed
2024-09-01 18:02:34 migration status error: failed - Unable to write to
socket: Broken pipe
2024-09-01 18:02:34 ERROR: online migrate failure - aborting
2024-09-01 18:02:34 aborting phase 2 - cleanup resources
2024-09-01 18:02:34 migrate_cancel
2024-09-01 18:02:36 ERROR: migration finished with problems (duration
00:00:07)
TASK ERROR: migration problems
```
If someone could point me in the correct direction to resolve this issue
I'd be very grateful - thanks
Cheer
Dulux-Oz
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PVE-User] VM Migration Not Happening :-(
2024-09-01 8:10 [PVE-User] VM Migration Not Happening :-( duluxoz
@ 2024-09-01 17:14 ` Lund Svein-Erik
2024-09-01 22:02 ` Gilberto Ferreira
1 sibling, 0 replies; 6+ messages in thread
From: Lund Svein-Erik @ 2024-09-01 17:14 UTC (permalink / raw)
To: matthew, Proxmox VE user list
I Can't give you any exact cause, but the message "channel 2: open failed....." comes from an issue with ssh between the nodes. I can't really tell anything else from the log you've posted.
---
---------------------------
From: "duluxoz"
To: "Proxmox VE user list"
Sent: Sunday, September 1, 2024 10:12 AM
Subject: [PVE-User] VM Migration Not Happening :-(
Hi All,
I need help with figuring out why I can't migrate a VM from one Proxmox
Node to another (in the same cluster, of course).
These are the details provided by the Proxmox Task Log:
```
task started by HA resource agent
2024-09-01 18:02:30 use dedicated network address for sending migration
traffic (192.168.200.103)
2024-09-01 18:02:30 starting migration of VM 100 to node 'pven3'
(192.168.200.103)
2024-09-01 18:02:30 starting VM 100 on remote node 'pven3'
2024-09-01 18:02:30 [pven3]
2024-09-01 18:02:32 start remote tunnel
2024-09-01 18:02:33 ssh tunnel ver 1
2024-09-01 18:02:33 starting online/live migration on
unix:/run/qemu-server/100.migrate
2024-09-01 18:02:33 set migration capabilities
2024-09-01 18:02:33 migration downtime limit: 100 ms
2024-09-01 18:02:33 migration cachesize: 256.0 MiB
2024-09-01 18:02:33 set migration parameters
2024-09-01 18:02:33 start migrate command to
unix:/run/qemu-server/100.migrate
channel 2: open failed: connect failed: open failed
2024-09-01 18:02:34 migration status error: failed - Unable to write to
socket: Broken pipe
2024-09-01 18:02:34 ERROR: online migrate failure - aborting
2024-09-01 18:02:34 aborting phase 2 - cleanup resources
2024-09-01 18:02:34 migrate_cancel
2024-09-01 18:02:36 ERROR: migration finished with problems (duration
00:00:07)
TASK ERROR: migration problems
```
If someone could point me in the correct direction to resolve this issue
I'd be very grateful - thanks
Cheer
Dulux-Oz
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PVE-User] VM Migration Not Happening :-(
2024-09-01 8:10 [PVE-User] VM Migration Not Happening :-( duluxoz
2024-09-01 17:14 ` Lund Svein-Erik
@ 2024-09-01 22:02 ` Gilberto Ferreira
2024-09-02 5:50 ` duluxoz
1 sibling, 1 reply; 6+ messages in thread
From: Gilberto Ferreira @ 2024-09-01 22:02 UTC (permalink / raw)
To: matthew, Proxmox VE user list
Hi
We need more details.
Send us the following:
cat /etc/pve/storage
cat /etc/pvedatacenter
cat /etc/pve/corosync.conf
cat /etc/hosts
cat /etc/network/interfaces
Can you ssh between the nodes?
---
Gilberto Nunes Ferreira
(47) 99676-7530 - Whatsapp / Telegram
Em dom., 1 de set. de 2024 às 05:11, duluxoz <duluxoz@gmail.com> escreveu:
> Hi All,
>
> I need help with figuring out why I can't migrate a VM from one Proxmox
> Node to another (in the same cluster, of course).
>
> These are the details provided by the Proxmox Task Log:
>
> ```
>
> task started by HA resource agent
> 2024-09-01 18:02:30 use dedicated network address for sending migration
> traffic (192.168.200.103)
> 2024-09-01 18:02:30 starting migration of VM 100 to node 'pven3'
> (192.168.200.103)
> 2024-09-01 18:02:30 starting VM 100 on remote node 'pven3'
> 2024-09-01 18:02:30 [pven3]
> 2024-09-01 18:02:32 start remote tunnel
> 2024-09-01 18:02:33 ssh tunnel ver 1
> 2024-09-01 18:02:33 starting online/live migration on
> unix:/run/qemu-server/100.migrate
> 2024-09-01 18:02:33 set migration capabilities
> 2024-09-01 18:02:33 migration downtime limit: 100 ms
> 2024-09-01 18:02:33 migration cachesize: 256.0 MiB
> 2024-09-01 18:02:33 set migration parameters
> 2024-09-01 18:02:33 start migrate command to
> unix:/run/qemu-server/100.migrate
> channel 2: open failed: connect failed: open failed
> 2024-09-01 18:02:34 migration status error: failed - Unable to write to
> socket: Broken pipe
> 2024-09-01 18:02:34 ERROR: online migrate failure - aborting
> 2024-09-01 18:02:34 aborting phase 2 - cleanup resources
> 2024-09-01 18:02:34 migrate_cancel
> 2024-09-01 18:02:36 ERROR: migration finished with problems (duration
> 00:00:07)
> TASK ERROR: migration problems
> ```
>
> If someone could point me in the correct direction to resolve this issue
> I'd be very grateful - thanks
>
> Cheer
>
> Dulux-Oz
>
>
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
>
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PVE-User] VM Migration Not Happening :-(
2024-09-01 22:02 ` Gilberto Ferreira
@ 2024-09-02 5:50 ` duluxoz
2024-09-02 7:51 ` Fabian Grünbichler
0 siblings, 1 reply; 6+ messages in thread
From: duluxoz @ 2024-09-02 5:50 UTC (permalink / raw)
To: Gilberto Ferreira, Proxmox VE user list
Hi Gilberto, and Thank you for getting back to me.
Just to be 100% clear: the Proxmox (with Hyper-Converged Ceph) Cluster
is working AOK, except for the face I can't migrate *any* of the VMs
(live or shutdown).
Yes, I can SSH into each node from every other node using:
* the hostname of the "management" NIC
* the hostname of the "migration traffic" NIC
* the ip address of the "managment" NIC
* the IP address of the "migration traffic" NIC
The VM's HDD is on the rbd storage (see below).
As requested:
/etc/pve/storage.cfg
```
dir: local
path /var/lib/vz
content vztmpl,iso,backup
lvmthin: local-lvm
thinpool data
vgname pve
content images,rootdir
rbd: rbd
content images,rootdir
krbd 0
pool rbd
cephfs: cephfs
path /data/cephfs
content backup,vztmpl,iso
fs-name cephfs
```
/etc/pve/datacenter.cfg
```
console: html5
crs: ha-rebalance-on-start=1
ha: shutdown_policy=migrate
keyboard: en-us
migration: secure,network=192.168.200.0/24
next-id: lower=1000
```
/etc/pve/corosync.conf
```
logging {
debug: off
to_syslog: yes
}
nodelist {
node {
name: pven1
nodeid: 1
quorum_votes: 1
ring0_addr: 192.168.100.101
ring1_addr: 192.168.200.101
}
node {
name: pven2
nodeid: 2
quorum_votes: 1
ring0_addr: 192.168.100.102
ring1_addr: 192.168.200.102
}
node {
name: pven3
nodeid: 3
quorum_votes: 1
ring0_addr: 192.168.100.103
ring1_addr: 192.168.200.103
}
}
quorum {
provider: corosync_votequorum
}
totem {
cluster_name: cluster1
config_version: 4
interface {
knet_link_priority: 10
linknumber: 0
}
interface {
knet_link_priority: 20
linknumber: 1
}
ip_version: ipv4-6
link_mode: passive
secauth: on
version: 2
}
```
/etc/hosts
```
127.0.0.1 localhost localhost.localdomain localhost4
localhost4.localdomain4
::1 localhost localhost.localdomain localhost6
localhost6.localdomain6 ip6-localhost ip6-loopback
192.168.100.101 pven1.mydomain.local pven1
192.168.100.102 pven2.mydomain.local pven2
192.168.100.103 pven3.mydomain.local pven3
192.168.200.101 pvent1.mydomain.local pvent1
192.168.200.102 pvent2.mydomain.local pvent2
192.168.200.103 pvent3.mydomain.local pvent3
```
/etc/network/interfaces (for pven1; pven2 & pven3 are the same, except
for the IP Address (see above))
```
auto lo
iface lo inet loopback
iface eno1 inet manual
iface eno2 inet manual
auto enp2s0
iface enp2s0 inet static
address 192.168.200.20/24
mtu 9000
auto bond0
iface bond0 inet manual
bond-slaves eno1 eno2
bond-mode 802.3ad
bond-xmit-hash-policy layer2+3
bond-miimon 100
bond-downdelay 200
bond-updelay 200
auto vmbr0
iface vmbr0 inet static
bridge-ports bond0
bridge-stp off
bridge-fd 0
bridge-vlan-aware yes
bridge-vids 100,200
auto vmbr0.200
iface vmbr0.200 inet static
address 192.168.100.101/24
gateway 192.168.100.1
```
Note: iface enp2s0 (on all 3 Nodes) sits on an isolated VLAN which
(obviously) has no gateway ie the only hosts on that VLAN are pven1,
pven2, & pven3, and all are "pingable" from each other.
Thanks for the help
Dulux-Oz
On 2/9/24 08:02, Gilberto Ferreira wrote:
> Hi
> We need more details.
> Send us the following:
>
> cat /etc/pve/storage
> cat /etc/pvedatacenter
> cat /etc/pve/corosync.conf
> cat /etc/hosts
> cat /etc/network/interfaces
>
> Can you ssh between the nodes?
>
>
> ---
>
>
> Gilberto Nunes Ferreira
> (47) 99676-7530 - Whatsapp / Telegram
>
>
>
>
>
>
> Em dom., 1 de set. de 2024 às 05:11, duluxoz <duluxoz@gmail.com> escreveu:
>
> Hi All,
>
> I need help with figuring out why I can't migrate a VM from one
> Proxmox
> Node to another (in the same cluster, of course).
>
> These are the details provided by the Proxmox Task Log:
>
> ```
>
> task started by HA resource agent
> 2024-09-01 18:02:30 use dedicated network address for sending
> migration
> traffic (192.168.200.103)
> 2024-09-01 18:02:30 starting migration of VM 100 to node 'pven3'
> (192.168.200.103)
> 2024-09-01 18:02:30 starting VM 100 on remote node 'pven3'
> 2024-09-01 18:02:30 [pven3]
> 2024-09-01 18:02:32 start remote tunnel
> 2024-09-01 18:02:33 ssh tunnel ver 1
> 2024-09-01 18:02:33 starting online/live migration on
> unix:/run/qemu-server/100.migrate
> 2024-09-01 18:02:33 set migration capabilities
> 2024-09-01 18:02:33 migration downtime limit: 100 ms
> 2024-09-01 18:02:33 migration cachesize: 256.0 MiB
> 2024-09-01 18:02:33 set migration parameters
> 2024-09-01 18:02:33 start migrate command to
> unix:/run/qemu-server/100.migrate
> channel 2: open failed: connect failed: open failed
> 2024-09-01 18:02:34 migration status error: failed - Unable to
> write to
> socket: Broken pipe
> 2024-09-01 18:02:34 ERROR: online migrate failure - aborting
> 2024-09-01 18:02:34 aborting phase 2 - cleanup resources
> 2024-09-01 18:02:34 migrate_cancel
> 2024-09-01 18:02:36 ERROR: migration finished with problems (duration
> 00:00:07)
> TASK ERROR: migration problems
> ```
>
> If someone could point me in the correct direction to resolve this
> issue
> I'd be very grateful - thanks
>
> Cheer
>
> Dulux-Oz
>
>
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PVE-User] VM Migration Not Happening :-(
2024-09-02 5:50 ` duluxoz
@ 2024-09-02 7:51 ` Fabian Grünbichler
2024-09-02 9:54 ` duluxoz
0 siblings, 1 reply; 6+ messages in thread
From: Fabian Grünbichler @ 2024-09-02 7:51 UTC (permalink / raw)
To: matthew, Proxmox VE user list, duluxoz, Gilberto Ferreira
> duluxoz <duluxoz@gmail.com> hat am 02.09.2024 07:50 CEST geschrieben:
>
>
> Hi Gilberto, and Thank you for getting back to me.
>
> Just to be 100% clear: the Proxmox (with Hyper-Converged Ceph) Cluster
> is working AOK, except for the face I can't migrate *any* of the VMs
> (live or shutdown).
>
> Yes, I can SSH into each node from every other node using:
>
> * the hostname of the "management" NIC
> * the hostname of the "migration traffic" NIC
> * the ip address of the "managment" NIC
> * the IP address of the "migration traffic" NIC
you probably configured SSH to restrict port/socket forwarding.
see `man sshd_config` for why this is usually not really a meaningful config change.
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PVE-User] VM Migration Not Happening :-(
2024-09-02 7:51 ` Fabian Grünbichler
@ 2024-09-02 9:54 ` duluxoz
0 siblings, 0 replies; 6+ messages in thread
From: duluxoz @ 2024-09-02 9:54 UTC (permalink / raw)
To: Fabian Grünbichler, matthew, Proxmox VE user list,
Gilberto Ferreira
Yeap - that was it
Thanks for helping me out 😬
Cheers
Dulux-Oz
On 2/9/24 17:51, Fabian Grünbichler wrote:
>> duluxoz <duluxoz@gmail.com> hat am 02.09.2024 07:50 CEST geschrieben:
>>
>>
>> Hi Gilberto, and Thank you for getting back to me.
>>
>> Just to be 100% clear: the Proxmox (with Hyper-Converged Ceph) Cluster
>> is working AOK, except for the face I can't migrate *any* of the VMs
>> (live or shutdown).
>>
>> Yes, I can SSH into each node from every other node using:
>>
>> * the hostname of the "management" NIC
>> * the hostname of the "migration traffic" NIC
>> * the ip address of the "managment" NIC
>> * the IP address of the "migration traffic" NIC
> you probably configured SSH to restrict port/socket forwarding.
>
> see `man sshd_config` for why this is usually not really a meaningful config change.
>
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2024-09-02 9:54 UTC | newest]
Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-09-01 8:10 [PVE-User] VM Migration Not Happening :-( duluxoz
2024-09-01 17:14 ` Lund Svein-Erik
2024-09-01 22:02 ` Gilberto Ferreira
2024-09-02 5:50 ` duluxoz
2024-09-02 7:51 ` Fabian Grünbichler
2024-09-02 9:54 ` duluxoz
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal