* [PVE-User] qm remote-migrate @ 2023-10-10 19:01 Fabrizio Cuseo 2023-10-11 7:32 ` Fabrizio Cuseo 2023-10-11 7:41 ` Stefan Lendl 0 siblings, 2 replies; 7+ messages in thread From: Fabrizio Cuseo @ 2023-10-10 19:01 UTC (permalink / raw) To: pve-user Hello. I am testing qm remote-migrate with 2 pve 8.0.4 clusters. Source cluster has one bridge with vlan id on every VM, destination cluster uses SDN and a different bridge (vnet) without vlanid. If I migrate the vm, i need to specify both bridge and vlan-id, but I have not found an option to do it. PS: after migration, on the new cluster the vm is running without any problem, but on source cluster remains locked and in migration, so I need to issue a "qm unlock vmid" and stop/delete it. I know that is an experimental feature, so I send my test results. Regards, Fabrizio -- --- Fabrizio Cuseo - mailto:f.cuseo@panservice.it ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-10 19:01 [PVE-User] qm remote-migrate Fabrizio Cuseo @ 2023-10-11 7:32 ` Fabrizio Cuseo 2023-10-11 7:41 ` Stefan Lendl 1 sibling, 0 replies; 7+ messages in thread From: Fabrizio Cuseo @ 2023-10-11 7:32 UTC (permalink / raw) To: Proxmox VE user list PS: I'm checking the qm.pm code, and I have found that I miss an (undocumented ?) option "delete", otherwise vm remains stopped on the source cluster (but "migrating locked"). I can't found any option to specify the vlan-tag mapping on destination cluster, where I am using sdn. Some comment ? Thanks, Fabrizio ----- Il 10-ott-23, alle 21:01, Fabrizio Cuseo f.cuseo@panservice.it ha scritto: > Hello. > I am testing qm remote-migrate with 2 pve 8.0.4 clusters. > Source cluster has one bridge with vlan id on every VM, destination cluster uses > SDN and a different bridge (vnet) without vlanid. > If I migrate the vm, i need to specify both bridge and vlan-id, but I have not > found an option to do it. > > PS: after migration, on the new cluster the vm is running without any problem, > but on source cluster remains locked and in migration, so I need to issue a "qm > unlock vmid" and stop/delete it. > > I know that is an experimental feature, so I send my test results. > > Regards, Fabrizio > > > -- > --- > Fabrizio Cuseo - mailto:f.cuseo@panservice.it > > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user -- --- Fabrizio Cuseo - mailto:f.cuseo@panservice.it ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-10 19:01 [PVE-User] qm remote-migrate Fabrizio Cuseo 2023-10-11 7:32 ` Fabrizio Cuseo @ 2023-10-11 7:41 ` Stefan Lendl 2023-10-11 7:55 ` Fabrizio Cuseo 1 sibling, 1 reply; 7+ messages in thread From: Stefan Lendl @ 2023-10-11 7:41 UTC (permalink / raw) To: Fabrizio Cuseo, Proxmox VE user list, pve-user Fabrizio Cuseo <f.cuseo@panservice.it> writes: Hello Fabrizio, To better understand your issue, the source cluster has a VM with a bridge with a VLAN tag assigned and the target cluster does not have the same setup but uses SDN (vnet) without vlan. After migration you manually changed the VMs configuration to match the new setup? What SDN configuration are you using on the traget cluster? Please send the output of the following: head -n -1 /etc/pve/sdn/*.cfg What was to exact command you ran to start the remote-migrate process? Did you notice any suspicios log messages in the source clusters journal? Usually I would ask you to send me the entire journal but this is not feasible on the mailing list. If necessary, I would recommend you open a Thread in our community forum and I will take a look there. https://forum.proxmox.com/ Best regards, Stefan Lendl > Hello. > I am testing qm remote-migrate with 2 pve 8.0.4 clusters. > Source cluster has one bridge with vlan id on every VM, destination cluster uses SDN and a different bridge (vnet) without vlanid. > If I migrate the vm, i need to specify both bridge and vlan-id, but I have not found an option to do it. > > PS: after migration, on the new cluster the vm is running without any problem, but on source cluster remains locked and in migration, so I need to issue a "qm unlock vmid" and stop/delete it. > > I know that is an experimental feature, so I send my test results. > > Regards, Fabrizio > > > -- > --- > Fabrizio Cuseo - mailto:f.cuseo@panservice.it > > _______________________________________________ > pve-user mailing list > pve-user@lists.proxmox.com > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-11 7:41 ` Stefan Lendl @ 2023-10-11 7:55 ` Fabrizio Cuseo 2023-10-11 9:37 ` Stefan Lendl 0 siblings, 1 reply; 7+ messages in thread From: Fabrizio Cuseo @ 2023-10-11 7:55 UTC (permalink / raw) To: Stefan Lendl; +Cc: Proxmox VE user list, pve-user ----- Il 11-ott-23, alle 9:41, Stefan Lendl s.lendl@proxmox.com ha scritto: > Fabrizio Cuseo <f.cuseo@panservice.it> writes: > > Hello Fabrizio, > > To better understand your issue, the source cluster has a VM with a > bridge with a VLAN tag assigned and the target cluster does not have the > same setup but uses SDN (vnet) without vlan. Yes, it's correct. > After migration you manually changed the VMs configuration to match the > new setup? I can't because remote-migration returns an error (I cannot specify vlan tag on that bridge) > What SDN configuration are you using on the traget cluster? > Please send the output of the following: > > head -n -1 /etc/pve/sdn/*.cfg Can I send you in private ? Because is full of customer's names :/ But, this is a part of files: ==> /etc/pve/sdn/zones.cfg <== vlan: ZonaVLAN bridge vmbr0 ipam pve qinq: VPVT bridge vmbr0 tag 929 ipam pve vlan-protocol 802.1q ==> /etc/pve/sdn/vnets.cfg <== vnet: test100 zone FWHous alias Vlan 100 Test 921 qinq tag 100 vnet: vlan902 zone ZonaVLAN alias Vlan 902 Private-Vlan tag 902 > What was to exact command you ran to start the remote-migrate process? qm remote-migrate 4980 4980 'host=172.16.20.41,apitoken=PVEAPIToken=root@pam!remotemigrate=hiddensecret,fingerprint=hiddenfingerprint' --target-bridge vlan902 --target-storage NfsMirror --online > Did you notice any suspicios log messages in the source clusters > journal? Source: tunnel: -> sending command "version" to remote tunnel: <- got reply 2023-10-10 18:08:48 local WS tunnel version: 2 2023-10-10 18:08:48 remote WS tunnel version: 2 2023-10-10 18:08:48 minimum required WS tunnel version: 2 websocket tunnel started 2023-10-10 18:08:48 starting migration of VM 4980 to node 'nodo01-cluster1' (172.16.20.41) tunnel: -> sending command "bwlimit" to remote tunnel: <- got reply 2023-10-10 18:08:49 found local disk 'CephCluster3Copie:vm-4980-disk-0' (attached) 2023-10-10 18:08:49 mapped: net0 from vmbr1 to vlan902 2023-10-10 18:08:49 Allocating volume for drive 'scsi0' on remote storage 'NfsMirror'.. tunnel: -> sending command "disk" to remote tunnel: <- got reply 2023-10-10 18:08:49 volume 'CephCluster3Copie:vm-4980-disk-0' is 'NfsMirror:4980/vm-4980-disk-0.raw' on the target tunnel: -> sending command "config" to remote tunnel: <- got reply tunnel: -> sending command "start" to remote tunnel: <- got reply 2023-10-10 18:08:50 ERROR: online migrate failure - error - tunnel command '{"start_params":{"forcemachine":"pc-i440fx-8.0+pve0","forcecpu":null,"statefile":"unix","skiplock":1},"cmd":"start","migrate_opts":{"network":null,"nbd ":{"scsi0":{"volid":"NfsMirror:4980/vm-4980-disk-0.raw","success":true,"drivestr":"NfsMirror:4980/vm-4980-disk-0.raw,discard=on,format=raw,size=64G"}},"nbd_proto_version":1,"storagemap":{"default":"NfsMirror"},"migratedfrom":"n ode06-cluster4","type":"websocket","remote_node":"nodo01-cluster1","spice_ticket":null}}' failed - failed to handle 'start' command - start failed: QEMU exited with code 1 2023-10-10 18:08:50 aborting phase 2 - cleanup resources 2023-10-10 18:08:50 migrate_cancel tunnel: -> sending command "stop" to remote tunnel: <- got reply tunnel: -> sending command "quit" to remote tunnel: <- got reply 2023-10-10 18:08:51 ERROR: migration finished with problems (duration 00:00:03) TASK ERROR: migration problems DESTINATION: mtunnel started received command 'version' received command 'bwlimit' received command 'disk' Formatting '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw', fmt=raw size=68719476736 preallocation=off received command 'config' update VM 4980: -agent 1 -boot order=scsi0;ide2;net0 -cores 2 -ide2 none,media=cdrom -memory 8192 -name SeafileProTestS3 -net0 e1000=86:64:73:AB:33:AE,bridge=vlan902,tag=902 -numa 1 -ostype l26 -scsi0 NfsMirror:4980 /vm-4980-disk-0.raw,discard=on,format=raw,size=64G -scsihw virtio-scsi-pci -smbios1 uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce -sockets 2 received command 'start' QEMU: vm vlans are not allowed on vnet vlan902 at /usr/share/perl5/PVE/Network/SDN/Zones/Plugin.pm line 228. QEMU: kvm: -netdev type=tap,id=net0,ifname=tap4980i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown: network script /var/lib/qemu-server/pve-bridge failed with status 6400 received command 'stop' received command 'quit' freeing volume 'NfsMirror:4980/vm-4980-disk-0.raw' as part of cleanup disk image '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw' does not exist switching to exit-mode, waiting for client to disconnect mtunnel exited TASK OK Source VM conf file: agent: 1 balloon: 2048 boot: order=scsi0;ide2;net0 cores: 2 ide2: none,media=cdrom memory: 4096 name: SeafileProTestS3 net0: virtio=86:64:73:AB:33:AE,bridge=vmbr1,tag=902 numa: 1 ostype: l26 scsi0: CephCluster3Copie:vm-4980-disk-0,discard=on,size=64G scsihw: virtio-scsi-pci smbios1: uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce sockets: 2 vmgenid: 035cd26d-c74e-405e-9b4d-481f26d9cf5f > Usually I would ask you to send me the entire journal but this is not > feasible on the mailing list. If necessary, I would recommend you open a > Thread in our community forum and I will take a look there. > https://forum.proxmox.com/ > > Best regards, > Stefan Lendl Thank you in advance, Fabrizio > >> Hello. >> I am testing qm remote-migrate with 2 pve 8.0.4 clusters. >> Source cluster has one bridge with vlan id on every VM, destination cluster uses >> SDN and a different bridge (vnet) without vlanid. >> If I migrate the vm, i need to specify both bridge and vlan-id, but I have not >> found an option to do it. >> >> PS: after migration, on the new cluster the vm is running without any problem, >> but on source cluster remains locked and in migration, so I need to issue a "qm >> unlock vmid" and stop/delete it. >> >> I know that is an experimental feature, so I send my test results. >> >> Regards, Fabrizio >> >> >> -- >> --- >> Fabrizio Cuseo - mailto:f.cuseo@panservice.it >> >> _______________________________________________ >> pve-user mailing list >> pve-user@lists.proxmox.com > > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user -- --- Fabrizio Cuseo - mailto:f.cuseo@panservice.it Direzione Generale - Panservice InterNetWorking Servizi Professionali per Internet ed il Networking Panservice e' associata AIIP - RIPE Local Registry Phone: +39 0773 410020 - Fax: +39 0773 470219 http://www.panservice.it mailto:info@panservice.it Numero verde nazionale: 800 901492 ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-11 7:55 ` Fabrizio Cuseo @ 2023-10-11 9:37 ` Stefan Lendl 2023-10-11 10:14 ` Fabrizio Cuseo 0 siblings, 1 reply; 7+ messages in thread From: Stefan Lendl @ 2023-10-11 9:37 UTC (permalink / raw) To: Fabrizio Cuseo; +Cc: Proxmox VE user list, pve-user Fabrizio Cuseo <f.cuseo@panservice.it> writes: Thanks for providing the details. I will investigate the situation and we will consider a solution for our upcoming SDN upgrade. As a solution for now, please try to remove the VLAN tag from the source VM and try to migrate again. The target net interface will not require a VLAN tag assigned (and is therefore not allowed) on the VM because it is configured via SDN already. Best regards, Stefan > ----- Il 11-ott-23, alle 9:41, Stefan Lendl s.lendl@proxmox.com ha scritto: > >> Fabrizio Cuseo <f.cuseo@panservice.it> writes: >> >> Hello Fabrizio, >> >> To better understand your issue, the source cluster has a VM with a >> bridge with a VLAN tag assigned and the target cluster does not have the >> same setup but uses SDN (vnet) without vlan. > > Yes, it's correct. > > >> After migration you manually changed the VMs configuration to match the >> new setup? > > I can't because remote-migration returns an error (I cannot specify vlan tag on that bridge) > >> What SDN configuration are you using on the traget cluster? >> Please send the output of the following: >> >> head -n -1 /etc/pve/sdn/*.cfg > > Can I send you in private ? Because is full of customer's names :/ > > But, this is a part of files: > > ==> /etc/pve/sdn/zones.cfg <== > vlan: ZonaVLAN > bridge vmbr0 > ipam pve > > qinq: VPVT > bridge vmbr0 > tag 929 > ipam pve > vlan-protocol 802.1q > > > ==> /etc/pve/sdn/vnets.cfg <== > vnet: test100 > zone FWHous > alias Vlan 100 Test 921 qinq > tag 100 > > vnet: vlan902 > zone ZonaVLAN > alias Vlan 902 Private-Vlan > tag 902 > > > > >> What was to exact command you ran to start the remote-migrate process? > > qm remote-migrate 4980 4980 'host=172.16.20.41,apitoken=PVEAPIToken=root@pam!remotemigrate=hiddensecret,fingerprint=hiddenfingerprint' --target-bridge vlan902 --target-storage NfsMirror --online > > > >> Did you notice any suspicios log messages in the source clusters >> journal? > > Source: > > tunnel: -> sending command "version" to remote > tunnel: <- got reply > 2023-10-10 18:08:48 local WS tunnel version: 2 > 2023-10-10 18:08:48 remote WS tunnel version: 2 > 2023-10-10 18:08:48 minimum required WS tunnel version: 2 > websocket tunnel started > 2023-10-10 18:08:48 starting migration of VM 4980 to node 'nodo01-cluster1' (172.16.20.41) > tunnel: -> sending command "bwlimit" to remote > tunnel: <- got reply > 2023-10-10 18:08:49 found local disk 'CephCluster3Copie:vm-4980-disk-0' (attached) > 2023-10-10 18:08:49 mapped: net0 from vmbr1 to vlan902 > 2023-10-10 18:08:49 Allocating volume for drive 'scsi0' on remote storage 'NfsMirror'.. > tunnel: -> sending command "disk" to remote > tunnel: <- got reply > 2023-10-10 18:08:49 volume 'CephCluster3Copie:vm-4980-disk-0' is 'NfsMirror:4980/vm-4980-disk-0.raw' on the target > tunnel: -> sending command "config" to remote > tunnel: <- got reply > tunnel: -> sending command "start" to remote > tunnel: <- got reply > 2023-10-10 18:08:50 ERROR: online migrate failure - error - tunnel command '{"start_params":{"forcemachine":"pc-i440fx-8.0+pve0","forcecpu":null,"statefile":"unix","skiplock":1},"cmd":"start","migrate_opts":{"network":null,"nbd > ":{"scsi0":{"volid":"NfsMirror:4980/vm-4980-disk-0.raw","success":true,"drivestr":"NfsMirror:4980/vm-4980-disk-0.raw,discard=on,format=raw,size=64G"}},"nbd_proto_version":1,"storagemap":{"default":"NfsMirror"},"migratedfrom":"n > ode06-cluster4","type":"websocket","remote_node":"nodo01-cluster1","spice_ticket":null}}' failed - failed to handle 'start' command - start failed: QEMU exited with code 1 > 2023-10-10 18:08:50 aborting phase 2 - cleanup resources > 2023-10-10 18:08:50 migrate_cancel > tunnel: -> sending command "stop" to remote > tunnel: <- got reply > tunnel: -> sending command "quit" to remote > tunnel: <- got reply > 2023-10-10 18:08:51 ERROR: migration finished with problems (duration 00:00:03) > > TASK ERROR: migration problems > > > > > > DESTINATION: > > mtunnel started > received command 'version' > received command 'bwlimit' > received command 'disk' > Formatting '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw', fmt=raw size=68719476736 preallocation=off > received command 'config' > update VM 4980: -agent 1 -boot order=scsi0;ide2;net0 -cores 2 -ide2 none,media=cdrom -memory 8192 -name SeafileProTestS3 -net0 e1000=86:64:73:AB:33:AE,bridge=vlan902,tag=902 -numa 1 -ostype l26 -scsi0 NfsMirror:4980 > /vm-4980-disk-0.raw,discard=on,format=raw,size=64G -scsihw virtio-scsi-pci -smbios1 uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce -sockets 2 > received command 'start' > QEMU: vm vlans are not allowed on vnet vlan902 at /usr/share/perl5/PVE/Network/SDN/Zones/Plugin.pm line 228. > QEMU: kvm: -netdev type=tap,id=net0,ifname=tap4980i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown: network script /var/lib/qemu-server/pve-bridge failed with status 6400 > received command 'stop' > received command 'quit' > freeing volume 'NfsMirror:4980/vm-4980-disk-0.raw' as part of cleanup > disk image '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw' does not exist > switching to exit-mode, waiting for client to disconnect > mtunnel exited > TASK OK > > > Source VM conf file: > > agent: 1 > balloon: 2048 > boot: order=scsi0;ide2;net0 > cores: 2 > ide2: none,media=cdrom > memory: 4096 > name: SeafileProTestS3 > net0: virtio=86:64:73:AB:33:AE,bridge=vmbr1,tag=902 > numa: 1 > ostype: l26 > scsi0: CephCluster3Copie:vm-4980-disk-0,discard=on,size=64G > scsihw: virtio-scsi-pci > smbios1: uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce > sockets: 2 > vmgenid: 035cd26d-c74e-405e-9b4d-481f26d9cf5f > > > > >> Usually I would ask you to send me the entire journal but this is not >> feasible on the mailing list. If necessary, I would recommend you open a >> Thread in our community forum and I will take a look there. > >> https://forum.proxmox.com/ >> >> Best regards, >> Stefan Lendl > > > Thank you in advance, Fabrizio > > >> >>> Hello. >>> I am testing qm remote-migrate with 2 pve 8.0.4 clusters. >>> Source cluster has one bridge with vlan id on every VM, destination cluster uses >>> SDN and a different bridge (vnet) without vlanid. >>> If I migrate the vm, i need to specify both bridge and vlan-id, but I have not >>> found an option to do it. >>> >>> PS: after migration, on the new cluster the vm is running without any problem, >>> but on source cluster remains locked and in migration, so I need to issue a "qm >>> unlock vmid" and stop/delete it. >>> >>> I know that is an experimental feature, so I send my test results. >>> >>> Regards, Fabrizio >>> >>> >>> -- >>> --- >>> Fabrizio Cuseo - mailto:f.cuseo@panservice.it >>> >>> _______________________________________________ >>> pve-user mailing list >>> pve-user@lists.proxmox.com >> > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user > > -- > --- > Fabrizio Cuseo - mailto:f.cuseo@panservice.it > Direzione Generale - Panservice InterNetWorking > Servizi Professionali per Internet ed il Networking > Panservice e' associata AIIP - RIPE Local Registry > Phone: +39 0773 410020 - Fax: +39 0773 470219 > http://www.panservice.it mailto:info@panservice.it > Numero verde nazionale: 800 901492 ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-11 9:37 ` Stefan Lendl @ 2023-10-11 10:14 ` Fabrizio Cuseo 2023-10-11 16:06 ` DERUMIER, Alexandre 0 siblings, 1 reply; 7+ messages in thread From: Fabrizio Cuseo @ 2023-10-11 10:14 UTC (permalink / raw) To: Stefan Lendl; +Cc: pve-user ----- Il 11-ott-23, alle 11:37, Stefan Lendl s.lendl@proxmox.com ha scritto: > Fabrizio Cuseo <f.cuseo@panservice.it> writes: > > Thanks for providing the details. > I will investigate the situation and we will consider a solution for our > upcoming SDN upgrade. > > As a solution for now, please try to remove the VLAN tag from the source VM > and try to migrate again. The target net interface will not require a > VLAN tag assigned (and is therefore not allowed) on the VM because it is > configured via SDN already. Yes, I have done with a test VM, but I can't do it with production VMs because if I remove vlan tag, the source VM will stop to work. But I can install and configure SDN on the source cluster (upgrading it to the last 8.x), create a vlan zone, a vnet with that vlan id, change source bridge to vnet bridge and removing the vlan tag, and migrate. (I have just tested and seems to work). Thank you again, Fabrizio > > Best regards, > Stefan > >> ----- Il 11-ott-23, alle 9:41, Stefan Lendl s.lendl@proxmox.com ha scritto: >> >>> Fabrizio Cuseo <f.cuseo@panservice.it> writes: >>> >>> Hello Fabrizio, >>> >>> To better understand your issue, the source cluster has a VM with a >>> bridge with a VLAN tag assigned and the target cluster does not have the >>> same setup but uses SDN (vnet) without vlan. >> >> Yes, it's correct. >> >> >>> After migration you manually changed the VMs configuration to match the >>> new setup? >> >> I can't because remote-migration returns an error (I cannot specify vlan tag on >> that bridge) >> >>> What SDN configuration are you using on the traget cluster? >>> Please send the output of the following: >>> >>> head -n -1 /etc/pve/sdn/*.cfg >> >> Can I send you in private ? Because is full of customer's names :/ >> >> But, this is a part of files: >> >> ==> /etc/pve/sdn/zones.cfg <== >> vlan: ZonaVLAN >> bridge vmbr0 >> ipam pve >> >> qinq: VPVT >> bridge vmbr0 >> tag 929 >> ipam pve >> vlan-protocol 802.1q >> >> >> ==> /etc/pve/sdn/vnets.cfg <== >> vnet: test100 >> zone FWHous >> alias Vlan 100 Test 921 qinq >> tag 100 >> >> vnet: vlan902 >> zone ZonaVLAN >> alias Vlan 902 Private-Vlan >> tag 902 >> >> >> >> >>> What was to exact command you ran to start the remote-migrate process? >> >> qm remote-migrate 4980 4980 >> 'host=172.16.20.41,apitoken=PVEAPIToken=root@pam!remotemigrate=hiddensecret,fingerprint=hiddenfingerprint' >> --target-bridge vlan902 --target-storage NfsMirror --online >> >> >> >>> Did you notice any suspicios log messages in the source clusters >>> journal? >> >> Source: >> >> tunnel: -> sending command "version" to remote >> tunnel: <- got reply >> 2023-10-10 18:08:48 local WS tunnel version: 2 >> 2023-10-10 18:08:48 remote WS tunnel version: 2 >> 2023-10-10 18:08:48 minimum required WS tunnel version: 2 >> websocket tunnel started >> 2023-10-10 18:08:48 starting migration of VM 4980 to node 'nodo01-cluster1' >> (172.16.20.41) >> tunnel: -> sending command "bwlimit" to remote >> tunnel: <- got reply >> 2023-10-10 18:08:49 found local disk 'CephCluster3Copie:vm-4980-disk-0' >> (attached) >> 2023-10-10 18:08:49 mapped: net0 from vmbr1 to vlan902 >> 2023-10-10 18:08:49 Allocating volume for drive 'scsi0' on remote storage >> 'NfsMirror'.. >> tunnel: -> sending command "disk" to remote >> tunnel: <- got reply >> 2023-10-10 18:08:49 volume 'CephCluster3Copie:vm-4980-disk-0' is >> 'NfsMirror:4980/vm-4980-disk-0.raw' on the target >> tunnel: -> sending command "config" to remote >> tunnel: <- got reply >> tunnel: -> sending command "start" to remote >> tunnel: <- got reply >> 2023-10-10 18:08:50 ERROR: online migrate failure - error - tunnel command >> '{"start_params":{"forcemachine":"pc-i440fx-8.0+pve0","forcecpu":null,"statefile":"unix","skiplock":1},"cmd":"start","migrate_opts":{"network":null,"nbd >> ":{"scsi0":{"volid":"NfsMirror:4980/vm-4980-disk-0.raw","success":true,"drivestr":"NfsMirror:4980/vm-4980-disk-0.raw,discard=on,format=raw,size=64G"}},"nbd_proto_version":1,"storagemap":{"default":"NfsMirror"},"migratedfrom":"n >> ode06-cluster4","type":"websocket","remote_node":"nodo01-cluster1","spice_ticket":null}}' >> failed - failed to handle 'start' command - start failed: QEMU exited with code >> 1 >> 2023-10-10 18:08:50 aborting phase 2 - cleanup resources >> 2023-10-10 18:08:50 migrate_cancel >> tunnel: -> sending command "stop" to remote >> tunnel: <- got reply >> tunnel: -> sending command "quit" to remote >> tunnel: <- got reply >> 2023-10-10 18:08:51 ERROR: migration finished with problems (duration 00:00:03) >> >> TASK ERROR: migration problems >> >> >> >> >> >> DESTINATION: >> >> mtunnel started >> received command 'version' >> received command 'bwlimit' >> received command 'disk' >> Formatting '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw', fmt=raw >> size=68719476736 preallocation=off >> received command 'config' >> update VM 4980: -agent 1 -boot order=scsi0;ide2;net0 -cores 2 -ide2 >> none,media=cdrom -memory 8192 -name SeafileProTestS3 -net0 >> e1000=86:64:73:AB:33:AE,bridge=vlan902,tag=902 -numa 1 -ostype l26 -scsi0 >> NfsMirror:4980 >> /vm-4980-disk-0.raw,discard=on,format=raw,size=64G -scsihw virtio-scsi-pci >> -smbios1 uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce -sockets 2 >> received command 'start' >> QEMU: vm vlans are not allowed on vnet vlan902 at >> /usr/share/perl5/PVE/Network/SDN/Zones/Plugin.pm line 228. >> QEMU: kvm: -netdev >> type=tap,id=net0,ifname=tap4980i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown: >> network script /var/lib/qemu-server/pve-bridge failed with status 6400 >> received command 'stop' >> received command 'quit' >> freeing volume 'NfsMirror:4980/vm-4980-disk-0.raw' as part of cleanup >> disk image '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw' does not exist >> switching to exit-mode, waiting for client to disconnect >> mtunnel exited >> TASK OK >> >> >> Source VM conf file: >> >> agent: 1 >> balloon: 2048 >> boot: order=scsi0;ide2;net0 >> cores: 2 >> ide2: none,media=cdrom >> memory: 4096 >> name: SeafileProTestS3 >> net0: virtio=86:64:73:AB:33:AE,bridge=vmbr1,tag=902 >> numa: 1 >> ostype: l26 >> scsi0: CephCluster3Copie:vm-4980-disk-0,discard=on,size=64G >> scsihw: virtio-scsi-pci >> smbios1: uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce >> sockets: 2 >> vmgenid: 035cd26d-c74e-405e-9b4d-481f26d9cf5f >> >> >> >> >>> Usually I would ask you to send me the entire journal but this is not >>> feasible on the mailing list. If necessary, I would recommend you open a >>> Thread in our community forum and I will take a look there. >> >>> https://forum.proxmox.com/ >>> >>> Best regards, >>> Stefan Lendl >> >> >> Thank you in advance, Fabrizio >> >> >>> >>>> Hello. >>>> I am testing qm remote-migrate with 2 pve 8.0.4 clusters. >>>> Source cluster has one bridge with vlan id on every VM, destination cluster uses >>>> SDN and a different bridge (vnet) without vlanid. >>>> If I migrate the vm, i need to specify both bridge and vlan-id, but I have not >>>> found an option to do it. >>>> >>>> PS: after migration, on the new cluster the vm is running without any problem, >>>> but on source cluster remains locked and in migration, so I need to issue a "qm >>>> unlock vmid" and stop/delete it. >>>> >>>> I know that is an experimental feature, so I send my test results. >>>> >>>> Regards, Fabrizio >>>> >>>> >>>> -- >>>> --- >>>> Fabrizio Cuseo - mailto:f.cuseo@panservice.it >>>> >>>> _______________________________________________ >>>> pve-user mailing list >>>> pve-user@lists.proxmox.com >>> > https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> >> -- >> --- >> Fabrizio Cuseo - mailto:f.cuseo@panservice.it >> Direzione Generale - Panservice InterNetWorking >> Servizi Professionali per Internet ed il Networking >> Panservice e' associata AIIP - RIPE Local Registry >> Phone: +39 0773 410020 - Fax: +39 0773 470219 >> http://www.panservice.it mailto:info@panservice.it > > Numero verde nazionale: 800 901492 -- --- Fabrizio Cuseo - mailto:f.cuseo@panservice.it Direzione Generale - Panservice InterNetWorking Servizi Professionali per Internet ed il Networking Panservice e' associata AIIP - RIPE Local Registry Phone: +39 0773 410020 - Fax: +39 0773 470219 http://www.panservice.it mailto:info@panservice.it Numero verde nazionale: 800 901492 ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PVE-User] qm remote-migrate 2023-10-11 10:14 ` Fabrizio Cuseo @ 2023-10-11 16:06 ` DERUMIER, Alexandre 0 siblings, 0 replies; 7+ messages in thread From: DERUMIER, Alexandre @ 2023-10-11 16:06 UTC (permalink / raw) To: s.lendl, f.cuseo, pve-user; +Cc: pve-user yes, currently they are no mapping for vlan, It should be easy to add in the code. I had similar case in production, migrating from vm tagged vlan to remote sdn vxlan. I finally first move to sdn without vlan on source cluster , then migrate to target cluster. But doing it directly in migrate could be better indeed :) -------- Message initial -------- De: Fabrizio Cuseo <f.cuseo@panservice.it> Répondre à: Fabrizio Cuseo <f.cuseo@panservice.it>, Proxmox VE user list <pve-user@lists.proxmox.com> À: Stefan Lendl <s.lendl@proxmox.com> Cc: pve-user <pve-user@pve.proxmox.com> Objet: Re: [PVE-User] qm remote-migrate Date: 11/10/2023 12:14:24 ----- Il 11-ott-23, alle 11:37, Stefan Lendl s.lendl@proxmox.com ha scritto: > Fabrizio Cuseo <f.cuseo@panservice.it> writes: > > Thanks for providing the details. > I will investigate the situation and we will consider a solution for > our > upcoming SDN upgrade. > > As a solution for now, please try to remove the VLAN tag from the > source VM > and try to migrate again. The target net interface will not require a > VLAN tag assigned (and is therefore not allowed) on the VM because it > is > configured via SDN already. Yes, I have done with a test VM, but I can't do it with production VMs because if I remove vlan tag, the source VM will stop to work. But I can install and configure SDN on the source cluster (upgrading it to the last 8.x), create a vlan zone, a vnet with that vlan id, change source bridge to vnet bridge and removing the vlan tag, and migrate. (I have just tested and seems to work). Thank you again, Fabrizio > > Best regards, > Stefan > > > ----- Il 11-ott-23, alle 9:41, Stefan Lendl s.lendl@proxmox.com ha > > scritto: > > > > > Fabrizio Cuseo <f.cuseo@panservice.it> writes: > > > > > > Hello Fabrizio, > > > > > > To better understand your issue, the source cluster has a VM with > > > a > > > bridge with a VLAN tag assigned and the target cluster does not > > > have the > > > same setup but uses SDN (vnet) without vlan. > > > > Yes, it's correct. > > > > > > > After migration you manually changed the VMs configuration to > > > match the > > > new setup? > > > > I can't because remote-migration returns an error (I cannot specify > > vlan tag on > > that bridge) > > > > > What SDN configuration are you using on the traget cluster? > > > Please send the output of the following: > > > > > > head -n -1 /etc/pve/sdn/*.cfg > > > > Can I send you in private ? Because is full of customer's names :/ > > > > But, this is a part of files: > > > > ==> /etc/pve/sdn/zones.cfg <== > > vlan: ZonaVLAN > > bridge vmbr0 > > ipam pve > > > > qinq: VPVT > > bridge vmbr0 > > tag 929 > > ipam pve > > vlan-protocol 802.1q > > > > > > ==> /etc/pve/sdn/vnets.cfg <== > > vnet: test100 > > zone FWHous > > alias Vlan 100 Test 921 qinq > > tag 100 > > > > vnet: vlan902 > > zone ZonaVLAN > > alias Vlan 902 Private-Vlan > > tag 902 > > > > > > > > > > > What was to exact command you ran to start the remote-migrate > > > process? > > > > qm remote-migrate 4980 4980 > > 'host=172.16.20.41,apitoken=PVEAPIToken=root@pam!remotemigrate=hidd > > ensecret,fingerprint=hiddenfingerprint' > > --target-bridge vlan902 --target-storage NfsMirror --online > > > > > > > > > Did you notice any suspicios log messages in the source clusters > > > journal? > > > > Source: > > > > tunnel: -> sending command "version" to remote > > tunnel: <- got reply > > 2023-10-10 18:08:48 local WS tunnel version: 2 > > 2023-10-10 18:08:48 remote WS tunnel version: 2 > > 2023-10-10 18:08:48 minimum required WS tunnel version: 2 > > websocket tunnel started > > 2023-10-10 18:08:48 starting migration of VM 4980 to node 'nodo01- > > cluster1' > > (172.16.20.41) > > tunnel: -> sending command "bwlimit" to remote > > tunnel: <- got reply > > 2023-10-10 18:08:49 found local disk 'CephCluster3Copie:vm-4980- > > disk-0' > > (attached) > > 2023-10-10 18:08:49 mapped: net0 from vmbr1 to vlan902 > > 2023-10-10 18:08:49 Allocating volume for drive 'scsi0' on remote > > storage > > 'NfsMirror'.. > > tunnel: -> sending command "disk" to remote > > tunnel: <- got reply > > 2023-10-10 18:08:49 volume 'CephCluster3Copie:vm-4980-disk-0' is > > 'NfsMirror:4980/vm-4980-disk-0.raw' on the target > > tunnel: -> sending command "config" to remote > > tunnel: <- got reply > > tunnel: -> sending command "start" to remote > > tunnel: <- got reply > > 2023-10-10 18:08:50 ERROR: online migrate failure - error - tunnel > > command > > '{"start_params":{"forcemachine":"pc-i440fx- > > 8.0+pve0","forcecpu":null,"statefile":"unix","skiplock":1},"cmd":"s > > tart","migrate_opts":{"network":null,"nbd > > ":{"scsi0":{"volid":"NfsMirror:4980/vm-4980-disk- > > 0.raw","success":true,"drivestr":"NfsMirror:4980/vm-4980-disk- > > 0.raw,discard=on,format=raw,size=64G"}},"nbd_proto_version":1,"stor > > agemap":{"default":"NfsMirror"},"migratedfrom":"n > > ode06-cluster4","type":"websocket","remote_node":"nodo01- > > cluster1","spice_ticket":null}}' > > failed - failed to handle 'start' command - start failed: QEMU > > exited with code > > 1 > > 2023-10-10 18:08:50 aborting phase 2 - cleanup resources > > 2023-10-10 18:08:50 migrate_cancel > > tunnel: -> sending command "stop" to remote > > tunnel: <- got reply > > tunnel: -> sending command "quit" to remote > > tunnel: <- got reply > > 2023-10-10 18:08:51 ERROR: migration finished with problems > > (duration 00:00:03) > > > > TASK ERROR: migration problems > > > > > > > > > > > > DESTINATION: > > > > mtunnel started > > received command 'version' > > received command 'bwlimit' > > received command 'disk' > > Formatting '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw', > > fmt=raw > > size=68719476736 preallocation=off > > received command 'config' > > update VM 4980: -agent 1 -boot order=scsi0;ide2;net0 -cores 2 -ide2 > > none,media=cdrom -memory 8192 -name SeafileProTestS3 -net0 > > e1000=86:64:73:AB:33:AE,bridge=vlan902,tag=902 -numa 1 -ostype l26 > > -scsi0 > > NfsMirror:4980 > > /vm-4980-disk-0.raw,discard=on,format=raw,size=64G -scsihw virtio- > > scsi-pci > > -smbios1 uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce -sockets 2 > > received command 'start' > > QEMU: vm vlans are not allowed on vnet vlan902 at > > /usr/share/perl5/PVE/Network/SDN/Zones/Plugin.pm line 228. > > QEMU: kvm: -netdev > > type=tap,id=net0,ifname=tap4980i0,script=/var/lib/qemu-server/pve- > > bridge,downscript=/var/lib/qemu-server/pve-bridgedown: > > network script /var/lib/qemu-server/pve-bridge failed with status > > 6400 > > received command 'stop' > > received command 'quit' > > freeing volume 'NfsMirror:4980/vm-4980-disk-0.raw' as part of > > cleanup > > disk image '/mnt/pve/NfsMirror/images/4980/vm-4980-disk-0.raw' does > > not exist > > switching to exit-mode, waiting for client to disconnect > > mtunnel exited > > TASK OK > > > > > > Source VM conf file: > > > > agent: 1 > > balloon: 2048 > > boot: order=scsi0;ide2;net0 > > cores: 2 > > ide2: none,media=cdrom > > memory: 4096 > > name: SeafileProTestS3 > > net0: virtio=86:64:73:AB:33:AE,bridge=vmbr1,tag=902 > > numa: 1 > > ostype: l26 > > scsi0: CephCluster3Copie:vm-4980-disk-0,discard=on,size=64G > > scsihw: virtio-scsi-pci > > smbios1: uuid=39a07e5b-16b5-45a3-aad9-4e3f2b4e87ce > > sockets: 2 > > vmgenid: 035cd26d-c74e-405e-9b4d-481f26d9cf5f > > > > > > > > > > > Usually I would ask you to send me the entire journal but this is > > > not > > > feasible on the mailing list. If necessary, I would recommend you > > > open a > > > Thread in our community forum and I will take a look there. > > > > > https://antiphishing.cetsi.fr/proxy/v3?i=SGI0YVJGNmxZNE90Z2thMFYL > > > WSxJOfIERJocpmb73Vs&r=SW5LV3JodE9QZkRVZ3JEYaKpfBJeBDlAX9E2aicRCRO > > > 3qsFIBX9zb4pDqGdxG45MOoGKkZ3R8w3DjSjAvqYgRg&f=bnJjU3hQT3pQSmNQZVE > > > 3aPE86c9l_NoW4_35ZkzMrkRr2MW_BjbnDTswuy0AwOve3AOvphs1mSlvmN7amrdG > > > SQ&u=https%3A//forum.proxmox.com/&k=dFBm > > > > > > Best regards, > > > Stefan Lendl > > > > > > Thank you in advance, Fabrizio > > > > > > > > > > > Hello. > > > > I am testing qm remote-migrate with 2 pve 8.0.4 clusters. > > > > Source cluster has one bridge with vlan id on every VM, > > > > destination cluster uses > > > > SDN and a different bridge (vnet) without vlanid. > > > > If I migrate the vm, i need to specify both bridge and vlan-id, > > > > but I have not > > > > found an option to do it. > > > > > > > > PS: after migration, on the new cluster the vm is running > > > > without any problem, > > > > but on source cluster remains locked and in migration, so I > > > > need to issue a "qm > > > > unlock vmid" and stop/delete it. > > > > > > > > I know that is an experimental feature, so I send my test > > > > results. > > > > > > > > Regards, Fabrizio > > > > > > > > > > > > -- > > > > --- > > > > Fabrizio Cuseo - mailto:f.cuseo@panservice.it > > > > > > > > _______________________________________________ > > > > pve-user mailing list > > > > pve-user@lists.proxmox.com > > > > https://antiphishing.cetsi.fr/proxy/v3?i=SGI0YVJGNmxZNE90Z2thMF > > > > YLWSxJOfIERJocpmb73Vs&r=SW5LV3JodE9QZkRVZ3JEYaKpfBJeBDlAX9E2aic > > > > RCRO3qsFIBX9zb4pDqGdxG45MOoGKkZ3R8w3DjSjAvqYgRg&f=bnJjU3hQT3pQS > > > > mNQZVE3aPE86c9l_NoW4_35ZkzMrkRr2MW_BjbnDTswuy0AwOve3AOvphs1mSlv > > > > mN7amrdGSQ&u=https%3A//lists.proxmox.com/cgi- > > > > bin/mailman/listinfo/pve-user&k=dFBm > > > > -- > > --- > > Fabrizio Cuseo - mailto:f.cuseo@panservice.it > > Direzione Generale - Panservice InterNetWorking > > Servizi Professionali per Internet ed il Networking > > Panservice e' associata AIIP - RIPE Local Registry > > Phone: +39 0773 410020 - Fax: +39 0773 470219 > > https://antiphishing.cetsi.fr/proxy/v3?i=SGI0YVJGNmxZNE90Z2thMFYLWS > > xJOfIERJocpmb73Vs&r=SW5LV3JodE9QZkRVZ3JEYaKpfBJeBDlAX9E2aicRCRO3qsF > > IBX9zb4pDqGdxG45MOoGKkZ3R8w3DjSjAvqYgRg&f=bnJjU3hQT3pQSmNQZVE3aPE86 > > c9l_NoW4_35ZkzMrkRr2MW_BjbnDTswuy0AwOve3AOvphs1mSlvmN7amrdGSQ&u=htt > > p%3A//www.panservice.it&k=dFBm mailto:info@panservice.it > > Numero verde nazionale: 800 901492 ^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2023-10-11 16:16 UTC | newest] Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2023-10-10 19:01 [PVE-User] qm remote-migrate Fabrizio Cuseo 2023-10-11 7:32 ` Fabrizio Cuseo 2023-10-11 7:41 ` Stefan Lendl 2023-10-11 7:55 ` Fabrizio Cuseo 2023-10-11 9:37 ` Stefan Lendl 2023-10-11 10:14 ` Fabrizio Cuseo 2023-10-11 16:06 ` DERUMIER, Alexandre
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox