public inbox for pve-user@lists.proxmox.com
 help / color / mirror / Atom feed
* [PVE-User] PVE 7 & LACP
@ 2021-07-21 15:17 Uwe Sauter
  2021-07-21 15:45 ` Gilberto Ferreira
  0 siblings, 1 reply; 7+ messages in thread
From: Uwe Sauter @ 2021-07-21 15:17 UTC (permalink / raw)
  To: pve-user

Hi all,

Fresh installation of PVE 7 with all available updates.

It looks like I'm unable to configure bonding of two interfaces using LCAP.

Output from "ip a":
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host
       valid_lft forever preferred_lft forever
2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master vmbr1224 state UP group default
qlen 1000
    link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
    altname enp24s0f0
3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default qlen 1000
    link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::207:43ff:fe18:8270/64 scope link
       valid_lft forever preferred_lft forever
4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default qlen 1000
    link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::207:43ff:fe18:8278/64 scope link
       valid_lft forever preferred_lft forever
5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
    link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
    altname enp24s0f1
6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
    link/infiniband 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
    link/infiniband 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000
    link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
    inet 10.224.203.1/16 scope global vmbr1224
       valid_lft forever preferred_lft forever
    inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
       valid_lft forever preferred_lft forever


My current /etc/network/interfaces, which is missing the bridges for the VMs:
auto lo
iface lo inet loopback

auto eno1
iface eno1 inet manual

auto enp59s0f4
iface enp59s0f4 inet manual
  mtu 9000

auto enp59s0f4d1
iface enp59s0f4d1 inet manual
  mtu 9000

auto bond0
iface bond0 inet manual
  slaves enp59s0f4 enp59s0f4d1
  bond-downdelay 200
  bond-miimon 100
  bond-mode 802.3ad
  bond-updelay 200

auto vmbr1224
iface vmbr1224 inet static
  address 10.224.203.1
  netmask 255.255.0.0
  gateway 10.224.200.1
  bridge_ports eno1
  bridge_stp off
  bridge_fd 0

##  # VLAN 155 nativ on bond0
##  auto vmbr155
##  iface vmbr155 inet manual
##    mtu 9000
##    bridge_ports bond0
##    bridge_stp off
##    bridge_fd 0


This config was copied over from the PVE 6 installation on the same host which experienced a SSD
failure and thus was re-installed with PVE 7.


Am I missing something? Am I being blind?


Any help is appreciated.


Thanks,

	Uwe






^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 15:17 [PVE-User] PVE 7 & LACP Uwe Sauter
@ 2021-07-21 15:45 ` Gilberto Ferreira
  2021-07-21 15:49   ` Uwe Sauter
  0 siblings, 1 reply; 7+ messages in thread
From: Gilberto Ferreira @ 2021-07-21 15:45 UTC (permalink / raw)
  To: uwe.sauter.de, Proxmox VE user list

Perhaps you need to change or add this feature in the network interface
file:

https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change



---
Gilberto Nunes Ferreira
(47) 99676-7530 - Whatsapp / Telegram






Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter <uwe.sauter.de@gmail.com>
escreveu:

> Hi all,
>
> Fresh installation of PVE 7 with all available updates.
>
> It looks like I'm unable to configure bonding of two interfaces using LCAP.
>
> Output from "ip a":
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
> default qlen 1000
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>        valid_lft forever preferred_lft forever
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
> vmbr1224 state UP group default
> qlen 1000
>     link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>     altname enp24s0f0
> 3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP
> group default qlen 1000
>     link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::207:43ff:fe18:8270/64 scope link
>        valid_lft forever preferred_lft forever
> 4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state
> UP group default qlen 1000
>     link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::207:43ff:fe18:8278/64 scope link
>        valid_lft forever preferred_lft forever
> 5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> default qlen 1000
>     link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
>     altname enp24s0f1
> 6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group
> default qlen 256
>     link/infiniband
> 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
> 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> 7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group
> default qlen 256
>     link/infiniband
> 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
> 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> 8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UP group default qlen 1000
>     link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>     inet 10.224.203.1/16 scope global vmbr1224
>        valid_lft forever preferred_lft forever
>     inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
>        valid_lft forever preferred_lft forever
>
>
> My current /etc/network/interfaces, which is missing the bridges for the
> VMs:
> auto lo
> iface lo inet loopback
>
> auto eno1
> iface eno1 inet manual
>
> auto enp59s0f4
> iface enp59s0f4 inet manual
>   mtu 9000
>
> auto enp59s0f4d1
> iface enp59s0f4d1 inet manual
>   mtu 9000
>
> auto bond0
> iface bond0 inet manual
>   slaves enp59s0f4 enp59s0f4d1
>   bond-downdelay 200
>   bond-miimon 100
>   bond-mode 802.3ad
>   bond-updelay 200
>
> auto vmbr1224
> iface vmbr1224 inet static
>   address 10.224.203.1
>   netmask 255.255.0.0
>   gateway 10.224.200.1
>   bridge_ports eno1
>   bridge_stp off
>   bridge_fd 0
>
> ##  # VLAN 155 nativ on bond0
> ##  auto vmbr155
> ##  iface vmbr155 inet manual
> ##    mtu 9000
> ##    bridge_ports bond0
> ##    bridge_stp off
> ##    bridge_fd 0
>
>
> This config was copied over from the PVE 6 installation on the same host
> which experienced a SSD
> failure and thus was re-installed with PVE 7.
>
>
> Am I missing something? Am I being blind?
>
>
> Any help is appreciated.
>
>
> Thanks,
>
>         Uwe
>
>
>
>
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
>


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 15:45 ` Gilberto Ferreira
@ 2021-07-21 15:49   ` Uwe Sauter
  2021-07-21 16:41     ` Gilberto Ferreira
  0 siblings, 1 reply; 7+ messages in thread
From: Uwe Sauter @ 2021-07-21 15:49 UTC (permalink / raw)
  To: Gilberto Ferreira, Proxmox VE user list

Hi Gilberto,

thanks for that tip. Unfortunately the network configuration quits a few steps before the bridge is
configured. The needed bond0 is not created…


Regards,

	Uwe


Am 21.07.21 um 17:45 schrieb Gilberto Ferreira:
> Perhaps you need to change or add this feature in the network interface file:
> 
> https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>
> 
> 
> 
> ---
> Gilberto Nunes Ferreira
> (47) 99676-7530 - Whatsapp / Telegram
> 
> 
> 
> 
> 
> 
> Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter <uwe.sauter.de@gmail.com
> <mailto:uwe.sauter.de@gmail.com>> escreveu:
> 
>     Hi all,
> 
>     Fresh installation of PVE 7 with all available updates.
> 
>     It looks like I'm unable to configure bonding of two interfaces using LCAP.
> 
>     Output from "ip a":
>     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
>         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>         inet 127.0.0.1/8 <http://127.0.0.1/8> scope host lo
>            valid_lft forever preferred_lft forever
>         inet6 ::1/128 scope host
>            valid_lft forever preferred_lft forever
>     2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master vmbr1224 state UP group default
>     qlen 1000
>         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>         altname enp24s0f0
>     3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default qlen 1000
>         link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
>         inet6 fe80::207:43ff:fe18:8270/64 scope link
>            valid_lft forever preferred_lft forever
>     4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default qlen 1000
>         link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
>         inet6 fe80::207:43ff:fe18:8278/64 scope link
>            valid_lft forever preferred_lft forever
>     5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
>         link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
>         altname enp24s0f1
>     6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>         link/infiniband 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
>     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>     7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>         link/infiniband 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
>     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>     8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default
>     qlen 1000
>         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>         inet 10.224.203.1/16 <http://10.224.203.1/16> scope global vmbr1224
>            valid_lft forever preferred_lft forever
>         inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
>            valid_lft forever preferred_lft forever
> 
> 
>     My current /etc/network/interfaces, which is missing the bridges for the VMs:
>     auto lo
>     iface lo inet loopback
> 
>     auto eno1
>     iface eno1 inet manual
> 
>     auto enp59s0f4
>     iface enp59s0f4 inet manual
>       mtu 9000
> 
>     auto enp59s0f4d1
>     iface enp59s0f4d1 inet manual
>       mtu 9000
> 
>     auto bond0
>     iface bond0 inet manual
>       slaves enp59s0f4 enp59s0f4d1
>       bond-downdelay 200
>       bond-miimon 100
>       bond-mode 802.3ad
>       bond-updelay 200
> 
>     auto vmbr1224
>     iface vmbr1224 inet static
>       address 10.224.203.1
>       netmask 255.255.0.0
>       gateway 10.224.200.1
>       bridge_ports eno1
>       bridge_stp off
>       bridge_fd 0
> 
>     ##  # VLAN 155 nativ on bond0
>     ##  auto vmbr155
>     ##  iface vmbr155 inet manual
>     ##    mtu 9000
>     ##    bridge_ports bond0
>     ##    bridge_stp off
>     ##    bridge_fd 0
> 
> 
>     This config was copied over from the PVE 6 installation on the same host which experienced a SSD
>     failure and thus was re-installed with PVE 7.
> 
> 
>     Am I missing something? Am I being blind?
> 
> 
>     Any help is appreciated.
> 
> 
>     Thanks,
> 
>             Uwe
> 
> 
> 
> 
>     _______________________________________________
>     pve-user mailing list
>     pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>
>     https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
> 




^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 15:49   ` Uwe Sauter
@ 2021-07-21 16:41     ` Gilberto Ferreira
  2021-07-21 17:31       ` Uwe Sauter
  0 siblings, 1 reply; 7+ messages in thread
From: Gilberto Ferreira @ 2021-07-21 16:41 UTC (permalink / raw)
  To: uwe.sauter.de; +Cc: Proxmox VE user list

So did you solve the issue?
---
Gilberto Nunes Ferreira
(47) 99676-7530 - Whatsapp / Telegram






Em qua., 21 de jul. de 2021 às 12:49, Uwe Sauter <uwe.sauter.de@gmail.com>
escreveu:

> Hi Gilberto,
>
> thanks for that tip. Unfortunately the network configuration quits a few
> steps before the bridge is
> configured. The needed bond0 is not created…
>
>
> Regards,
>
>         Uwe
>
>
> Am 21.07.21 um 17:45 schrieb Gilberto Ferreira:
> > Perhaps you need to change or add this feature in the network interface
> file:
> >
> >
> https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> > <
> https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> >
> >
> >
> >
> > ---
> > Gilberto Nunes Ferreira
> > (47) 99676-7530 - Whatsapp / Telegram
> >
> >
> >
> >
> >
> >
> > Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter <
> uwe.sauter.de@gmail.com
> > <mailto:uwe.sauter.de@gmail.com>> escreveu:
> >
> >     Hi all,
> >
> >     Fresh installation of PVE 7 with all available updates.
> >
> >     It looks like I'm unable to configure bonding of two interfaces
> using LCAP.
> >
> >     Output from "ip a":
> >     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> group default qlen 1000
> >         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >         inet 127.0.0.1/8 <http://127.0.0.1/8> scope host lo
> >            valid_lft forever preferred_lft forever
> >         inet6 ::1/128 scope host
> >            valid_lft forever preferred_lft forever
> >     2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
> vmbr1224 state UP group default
> >     qlen 1000
> >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
> >         altname enp24s0f0
> >     3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq
> state UP group default qlen 1000
> >         link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
> >         inet6 fe80::207:43ff:fe18:8270/64 scope link
> >            valid_lft forever preferred_lft forever
> >     4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq
> state UP group default qlen 1000
> >         link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
> >         inet6 fe80::207:43ff:fe18:8278/64 scope link
> >            valid_lft forever preferred_lft forever
> >     5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> default qlen 1000
> >         link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
> >         altname enp24s0f1
> >     6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN
> group default qlen 256
> >         link/infiniband
> 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
> >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> >     7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN
> group default qlen 256
> >         link/infiniband
> 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
> >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> >     8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP group default
> >     qlen 1000
> >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
> >         inet 10.224.203.1/16 <http://10.224.203.1/16> scope global
> vmbr1224
> >            valid_lft forever preferred_lft forever
> >         inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
> >            valid_lft forever preferred_lft forever
> >
> >
> >     My current /etc/network/interfaces, which is missing the bridges for
> the VMs:
> >     auto lo
> >     iface lo inet loopback
> >
> >     auto eno1
> >     iface eno1 inet manual
> >
> >     auto enp59s0f4
> >     iface enp59s0f4 inet manual
> >       mtu 9000
> >
> >     auto enp59s0f4d1
> >     iface enp59s0f4d1 inet manual
> >       mtu 9000
> >
> >     auto bond0
> >     iface bond0 inet manual
> >       slaves enp59s0f4 enp59s0f4d1
> >       bond-downdelay 200
> >       bond-miimon 100
> >       bond-mode 802.3ad
> >       bond-updelay 200
> >
> >     auto vmbr1224
> >     iface vmbr1224 inet static
> >       address 10.224.203.1
> >       netmask 255.255.0.0
> >       gateway 10.224.200.1
> >       bridge_ports eno1
> >       bridge_stp off
> >       bridge_fd 0
> >
> >     ##  # VLAN 155 nativ on bond0
> >     ##  auto vmbr155
> >     ##  iface vmbr155 inet manual
> >     ##    mtu 9000
> >     ##    bridge_ports bond0
> >     ##    bridge_stp off
> >     ##    bridge_fd 0
> >
> >
> >     This config was copied over from the PVE 6 installation on the same
> host which experienced a SSD
> >     failure and thus was re-installed with PVE 7.
> >
> >
> >     Am I missing something? Am I being blind?
> >
> >
> >     Any help is appreciated.
> >
> >
> >     Thanks,
> >
> >             Uwe
> >
> >
> >
> >
> >     _______________________________________________
> >     pve-user mailing list
> >     pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>
> >     https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
> >
>
>


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 16:41     ` Gilberto Ferreira
@ 2021-07-21 17:31       ` Uwe Sauter
  2021-07-21 17:58         ` alexandre derumier
  0 siblings, 1 reply; 7+ messages in thread
From: Uwe Sauter @ 2021-07-21 17:31 UTC (permalink / raw)
  To: Gilberto Ferreira; +Cc: Proxmox VE user list

No, the bond isn't created on system boot. Thus the bridge that should connect the bond with the VMs
is created but the connection is not made.


Am 21.07.21 um 18:41 schrieb Gilberto Ferreira:
> So did you solve the issue?
> ---
> Gilberto Nunes Ferreira
> (47) 99676-7530 - Whatsapp / Telegram
> 
> 
> 
> 
> 
> 
> Em qua., 21 de jul. de 2021 às 12:49, Uwe Sauter <uwe.sauter.de@gmail.com
> <mailto:uwe.sauter.de@gmail.com>> escreveu:
> 
>     Hi Gilberto,
> 
>     thanks for that tip. Unfortunately the network configuration quits a few steps before the bridge is
>     configured. The needed bond0 is not created…
> 
> 
>     Regards,
> 
>             Uwe
> 
> 
>     Am 21.07.21 um 17:45 schrieb Gilberto Ferreira:
>     > Perhaps you need to change or add this feature in the network interface file:
>     >
>     > https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>     <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>
>     > <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>     <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>>
>     >
>     >
>     >
>     > ---
>     > Gilberto Nunes Ferreira
>     > (47) 99676-7530 - Whatsapp / Telegram
>     >
>     >
>     >
>     >
>     >
>     >
>     > Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter <uwe.sauter.de@gmail.com
>     <mailto:uwe.sauter.de@gmail.com>
>     > <mailto:uwe.sauter.de@gmail.com <mailto:uwe.sauter.de@gmail.com>>> escreveu:
>     >
>     >     Hi all,
>     >
>     >     Fresh installation of PVE 7 with all available updates.
>     >
>     >     It looks like I'm unable to configure bonding of two interfaces using LCAP.
>     >
>     >     Output from "ip a":
>     >     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
>     >         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     >         inet 127.0.0.1/8 <http://127.0.0.1/8> <http://127.0.0.1/8 <http://127.0.0.1/8>> scope
>     host lo
>     >            valid_lft forever preferred_lft forever
>     >         inet6 ::1/128 scope host
>     >            valid_lft forever preferred_lft forever
>     >     2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master vmbr1224 state UP
>     group default
>     >     qlen 1000
>     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>     >         altname enp24s0f0
>     >     3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default
>     qlen 1000
>     >         link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
>     >         inet6 fe80::207:43ff:fe18:8270/64 scope link
>     >            valid_lft forever preferred_lft forever
>     >     4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default
>     qlen 1000
>     >         link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
>     >         inet6 fe80::207:43ff:fe18:8278/64 scope link
>     >            valid_lft forever preferred_lft forever
>     >     5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
>     >         link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
>     >         altname enp24s0f1
>     >     6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>     >         link/infiniband 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
>     >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>     >     7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>     >         link/infiniband 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
>     >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>     >     8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default
>     >     qlen 1000
>     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>     >         inet 10.224.203.1/16 <http://10.224.203.1/16> <http://10.224.203.1/16
>     <http://10.224.203.1/16>> scope global vmbr1224
>     >            valid_lft forever preferred_lft forever
>     >         inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
>     >            valid_lft forever preferred_lft forever
>     >
>     >
>     >     My current /etc/network/interfaces, which is missing the bridges for the VMs:
>     >     auto lo
>     >     iface lo inet loopback
>     >
>     >     auto eno1
>     >     iface eno1 inet manual
>     >
>     >     auto enp59s0f4
>     >     iface enp59s0f4 inet manual
>     >       mtu 9000
>     >
>     >     auto enp59s0f4d1
>     >     iface enp59s0f4d1 inet manual
>     >       mtu 9000
>     >
>     >     auto bond0
>     >     iface bond0 inet manual
>     >       slaves enp59s0f4 enp59s0f4d1
>     >       bond-downdelay 200
>     >       bond-miimon 100
>     >       bond-mode 802.3ad
>     >       bond-updelay 200
>     >
>     >     auto vmbr1224
>     >     iface vmbr1224 inet static
>     >       address 10.224.203.1
>     >       netmask 255.255.0.0
>     >       gateway 10.224.200.1
>     >       bridge_ports eno1
>     >       bridge_stp off
>     >       bridge_fd 0
>     >
>     >     ##  # VLAN 155 nativ on bond0
>     >     ##  auto vmbr155
>     >     ##  iface vmbr155 inet manual
>     >     ##    mtu 9000
>     >     ##    bridge_ports bond0
>     >     ##    bridge_stp off
>     >     ##    bridge_fd 0
>     >
>     >
>     >     This config was copied over from the PVE 6 installation on the same host which experienced
>     a SSD
>     >     failure and thus was re-installed with PVE 7.
>     >
>     >
>     >     Am I missing something? Am I being blind?
>     >
>     >
>     >     Any help is appreciated.
>     >
>     >
>     >     Thanks,
>     >
>     >             Uwe
>     >
>     >
>     >
>     >
>     >     _______________________________________________
>     >     pve-user mailing list
>     >     pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>
>     <mailto:pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>>
>     >     https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
>     >     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>>
>     >
> 




^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 17:31       ` Uwe Sauter
@ 2021-07-21 17:58         ` alexandre derumier
  2021-07-21 18:05           ` Uwe Sauter
  0 siblings, 1 reply; 7+ messages in thread
From: alexandre derumier @ 2021-07-21 17:58 UTC (permalink / raw)
  To: uwe.sauter.de, Proxmox VE user list, Gilberto Ferreira

slaves enp59s0f4 enp59s0f4d1

---->

bond-slaves enp59s0f4 enp59s0f4d1


("slaves" is deprecated since years, and ifupdown2 don't support it)


Le mercredi 21 juillet 2021 à 19:31 +0200, Uwe Sauter a écrit :
> No, the bond isn't created on system boot. Thus the bridge that
> should connect the bond with the VMs
> is created but the connection is not made.
> 
> 
> Am 21.07.21 um 18:41 schrieb Gilberto Ferreira:
> > So did you solve the issue?
> > ---
> > Gilberto Nunes Ferreira
> > (47) 99676-7530 - Whatsapp / Telegram
> > 
> > 
> > 
> > 
> > 
> > 
> > Em qua., 21 de jul. de 2021 às 12:49, Uwe Sauter
> > <uwe.sauter.de@gmail.com
> > <mailto:uwe.sauter.de@gmail.com>> escreveu:
> > 
> >     Hi Gilberto,
> > 
> >     thanks for that tip. Unfortunately the network configuration
> > quits a few steps before the bridge is
> >     configured. The needed bond0 is not created…
> > 
> > 
> >     Regards,
> > 
> >             Uwe
> > 
> > 
> >     Am 21.07.21 um 17:45 schrieb Gilberto Ferreira:
> >     > Perhaps you need to change or add this feature in the network
> > interface file:
> >     >
> >     >
> > https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> >    
> > <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> > >
> >     >
> > <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> >    
> > <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
> > >>
> >     >
> >     >
> >     >
> >     > ---
> >     > Gilberto Nunes Ferreira
> >     > (47) 99676-7530 - Whatsapp / Telegram
> >     >
> >     >
> >     >
> >     >
> >     >
> >     >
> >     > Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter
> > <uwe.sauter.de@gmail.com
> >     <mailto:uwe.sauter.de@gmail.com>
> >     >
> > <mailto:uwe.sauter.de@gmail.com <mailto:uwe.sauter.de@gmail.com>>>
> > escreveu:
> >     >
> >     >     Hi all,
> >     >
> >     >     Fresh installation of PVE 7 with all available updates.
> >     >
> >     >     It looks like I'm unable to configure bonding of two
> > interfaces using LCAP.
> >     >
> >     >     Output from "ip a":
> >     >     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue
> > state UNKNOWN group default qlen 1000
> >     >         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >     >         inet 127.0.0.1/8 <http://127.0.0.1/8>
> > <http://127.0.0.1/8 <http://127.0.0.1/8>> scope
> >     host lo
> >     >            valid_lft forever preferred_lft forever
> >     >         inet6 ::1/128 scope host
> >     >            valid_lft forever preferred_lft forever
> >     >     2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> > mq master vmbr1224 state UP
> >     group default
> >     >     qlen 1000
> >     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
> >     >         altname enp24s0f0
> >     >     3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000
> > qdisc mq state UP group default
> >     qlen 1000
> >     >         link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
> >     >         inet6 fe80::207:43ff:fe18:8270/64 scope link
> >     >            valid_lft forever preferred_lft forever
> >     >     4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> > 9000 qdisc mq state UP group default
> >     qlen 1000
> >     >         link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
> >     >         inet6 fe80::207:43ff:fe18:8278/64 scope link
> >     >            valid_lft forever preferred_lft forever
> >     >     5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state
> > DOWN group default qlen 1000
> >     >         link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
> >     >         altname enp24s0f1
> >     >     6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop
> > state DOWN group default qlen 256
> >     >         link/infiniband
> > 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
> >     >   
> >  00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> >     >     7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop
> > state DOWN group default qlen 256
> >     >         link/infiniband
> > 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
> >     >   
> >  00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> >     >     8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> > qdisc noqueue state UP group default
> >     >     qlen 1000
> >     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
> >     >         inet 10.224.203.1/16 <http://10.224.203.1/16>
> > <http://10.224.203.1/16
> >     <http://10.224.203.1/16>> scope global vmbr1224
> >     >            valid_lft forever preferred_lft forever
> >     >         inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
> >     >            valid_lft forever preferred_lft forever
> >     >
> >     >
> >     >     My current /etc/network/interfaces, which is missing the
> > bridges for the VMs:
> >     >     auto lo
> >     >     iface lo inet loopback
> >     >
> >     >     auto eno1
> >     >     iface eno1 inet manual
> >     >
> >     >     auto enp59s0f4
> >     >     iface enp59s0f4 inet manual
> >     >       mtu 9000
> >     >
> >     >     auto enp59s0f4d1
> >     >     iface enp59s0f4d1 inet manual
> >     >       mtu 9000
> >     >
> >     >     auto bond0
> >     >     iface bond0 inet manual
> >     >       slaves enp59s0f4 enp59s0f4d1
> >     >       bond-downdelay 200
> >     >       bond-miimon 100
> >     >       bond-mode 802.3ad
> >     >       bond-updelay 200
> >     >
> >     >     auto vmbr1224
> >     >     iface vmbr1224 inet static
> >     >       address 10.224.203.1
> >     >       netmask 255.255.0.0
> >     >       gateway 10.224.200.1
> >     >       bridge_ports eno1
> >     >       bridge_stp off
> >     >       bridge_fd 0
> >     >
> >     >     ##  # VLAN 155 nativ on bond0
> >     >     ##  auto vmbr155
> >     >     ##  iface vmbr155 inet manual
> >     >     ##    mtu 9000
> >     >     ##    bridge_ports bond0
> >     >     ##    bridge_stp off
> >     >     ##    bridge_fd 0
> >     >
> >     >
> >     >     This config was copied over from the PVE 6 installation
> > on the same host which experienced
> >     a SSD
> >     >     failure and thus was re-installed with PVE 7.
> >     >
> >     >
> >     >     Am I missing something? Am I being blind?
> >     >
> >     >
> >     >     Any help is appreciated.
> >     >
> >     >
> >     >     Thanks,
> >     >
> >     >             Uwe
> >     >
> >     >
> >     >
> >     >
> >     >     _______________________________________________
> >     >     pve-user mailing list
> >     >   
> >  pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>
> >    
> > <mailto:pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com
> > >>
> >     >   
> >  https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
> >     >   
> >  <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>>
> >     >
> > 
> 
> 
> _______________________________________________
> pve-user mailing list
> pve-user@lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user



^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PVE-User] PVE 7 & LACP
  2021-07-21 17:58         ` alexandre derumier
@ 2021-07-21 18:05           ` Uwe Sauter
  0 siblings, 0 replies; 7+ messages in thread
From: Uwe Sauter @ 2021-07-21 18:05 UTC (permalink / raw)
  To: alexandre derumier, Proxmox VE user list, Gilberto Ferreira

Am 21.07.21 um 19:58 schrieb alexandre derumier:
> slaves enp59s0f4 enp59s0f4d1
> 
> ---->
> 
> bond-slaves enp59s0f4 enp59s0f4d1

Thank you, that's it.

> 
> 
> ("slaves" is deprecated since years, and ifupdown2 don't support it)

Well, even the documentation for Buster still mentions "slaves" so it cannot be deprecated for that
long.

Regards,

	Uwe



> 
> 
> Le mercredi 21 juillet 2021 à 19:31 +0200, Uwe Sauter a écrit :
>> No, the bond isn't created on system boot. Thus the bridge that should connect the bond with the VMs
>> is created but the connection is not made.
>>
>>
>> Am 21.07.21 um 18:41 schrieb Gilberto Ferreira:
>>> So did you solve the issue?
>>> ---
>>> Gilberto Nunes Ferreira
>>> (47) 99676-7530 - Whatsapp / Telegram
>>>
>>>
>>>
>>>
>>>
>>>
>>> Em qua., 21 de jul. de 2021 às 12:49, Uwe Sauter <uwe.sauter.de@gmail.com
>>> <mailto:uwe.sauter.de@gmail.com>
>>> <mailto:uwe.sauter.de@gmail.com <mailto:uwe.sauter.de@gmail.com>>> escreveu:
>>>
>>>     Hi Gilberto,
>>>
>>>     thanks for that tip. Unfortunately the network configuration quits a few steps before the
>>> bridge is
>>>     configured. The needed bond0 is not created…
>>>
>>>
>>>     Regards,
>>>
>>>             Uwe
>>>
>>>
>>>     Am 21.07.21 um 17:45 schrieb Gilberto Ferreira:
>>>     > Perhaps you need to change or add this feature in the network interface file:
>>>     >
>>>     > https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>>> <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>
>>>     <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>>> <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>>
>>>     > <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>>> <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>
>>>     <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change
>>> <https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0#Linux_Bridge_MAC-Address_Change>>>
>>>     >
>>>     >
>>>     >
>>>     > ---
>>>     > Gilberto Nunes Ferreira
>>>     > (47) 99676-7530 - Whatsapp / Telegram
>>>     >
>>>     >
>>>     >
>>>     >
>>>     >
>>>     >
>>>     > Em qua., 21 de jul. de 2021 às 12:18, Uwe Sauter <uwe.sauter.de@gmail.com
>>> <mailto:uwe.sauter.de@gmail.com>
>>>     <mailto:uwe.sauter.de@gmail.com <mailto:uwe.sauter.de@gmail.com>>
>>>     > <mailto:uwe.sauter.de@gmail.com
>>> <mailto:uwe.sauter.de@gmail.com> <mailto:uwe.sauter.de@gmail.com
>>> <mailto:uwe.sauter.de@gmail.com>>>> escreveu:
>>>     >
>>>     >     Hi all,
>>>     >
>>>     >     Fresh installation of PVE 7 with all available updates.
>>>     >
>>>     >     It looks like I'm unable to configure bonding of two interfaces using LCAP.
>>>     >
>>>     >     Output from "ip a":
>>>     >     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
>>>     >         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>     >         inet 127.0.0.1/8 <http://127.0.0.1/8 <http://127.0.0.1/8>> <http://127.0.0.1/8
>>> <http://127.0.0.1/8> <http://127.0.0.1/8 <http://127.0.0.1/8>>> scope
>>>     host lo
>>>     >            valid_lft forever preferred_lft forever
>>>     >         inet6 ::1/128 scope host
>>>     >            valid_lft forever preferred_lft forever
>>>     >     2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master vmbr1224 state UP
>>>     group default
>>>     >     qlen 1000
>>>     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>>>     >         altname enp24s0f0
>>>     >     3: enp59s0f4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default
>>>     qlen 1000
>>>     >         link/ether 00:07:43:18:82:70 brd ff:ff:ff:ff:ff:ff
>>>     >         inet6 fe80::207:43ff:fe18:8270/64 scope link
>>>     >            valid_lft forever preferred_lft forever
>>>     >     4: enp59s0f4d1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP group default
>>>     qlen 1000
>>>     >         link/ether 00:07:43:18:82:78 brd ff:ff:ff:ff:ff:ff
>>>     >         inet6 fe80::207:43ff:fe18:8278/64 scope link
>>>     >            valid_lft forever preferred_lft forever
>>>     >     5: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
>>>     >         link/ether a4:bf:01:2f:63:f3 brd ff:ff:ff:ff:ff:ff
>>>     >         altname enp24s0f1
>>>     >     6: ibp94s0: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>>>     >         link/infiniband 80:00:0a:08:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:51 brd
>>>     >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>>>     >     7: ibp94s0d1: <BROADCAST,MULTICAST> mtu 2044 qdisc noop state DOWN group default qlen 256
>>>     >         link/infiniband 80:00:0a:09:fe:80:00:00:00:00:00:00:f4:52:14:03:00:2c:14:52 brd
>>>     >     00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
>>>     >     8: vmbr1224: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group
>>> default
>>>     >     qlen 1000
>>>     >         link/ether a4:bf:01:2f:63:f2 brd ff:ff:ff:ff:ff:ff
>>>     >         inet 10.224.203.1/16 <http://10.224.203.1/16 <http://10.224.203.1/16>>
>>> <http://10.224.203.1/16 <http://10.224.203.1/16>
>>>     <http://10.224.203.1/16 <http://10.224.203.1/16>>> scope global vmbr1224
>>>     >            valid_lft forever preferred_lft forever
>>>     >         inet6 fe80::a6bf:1ff:fe2f:63f2/64 scope link
>>>     >            valid_lft forever preferred_lft forever
>>>     >
>>>     >
>>>     >     My current /etc/network/interfaces, which is missing the bridges for the VMs:
>>>     >     auto lo
>>>     >     iface lo inet loopback
>>>     >
>>>     >     auto eno1
>>>     >     iface eno1 inet manual
>>>     >
>>>     >     auto enp59s0f4
>>>     >     iface enp59s0f4 inet manual
>>>     >       mtu 9000
>>>     >
>>>     >     auto enp59s0f4d1
>>>     >     iface enp59s0f4d1 inet manual
>>>     >       mtu 9000
>>>     >
>>>     >     auto bond0
>>>     >     iface bond0 inet manual
>>>     >       slaves enp59s0f4 enp59s0f4d1
>>>     >       bond-downdelay 200
>>>     >       bond-miimon 100
>>>     >       bond-mode 802.3ad
>>>     >       bond-updelay 200
>>>     >
>>>     >     auto vmbr1224
>>>     >     iface vmbr1224 inet static
>>>     >       address 10.224.203.1
>>>     >       netmask 255.255.0.0
>>>     >       gateway 10.224.200.1
>>>     >       bridge_ports eno1
>>>     >       bridge_stp off
>>>     >       bridge_fd 0
>>>     >
>>>     >     ##  # VLAN 155 nativ on bond0
>>>     >     ##  auto vmbr155
>>>     >     ##  iface vmbr155 inet manual
>>>     >     ##    mtu 9000
>>>     >     ##    bridge_ports bond0
>>>     >     ##    bridge_stp off
>>>     >     ##    bridge_fd 0
>>>     >
>>>     >
>>>     >     This config was copied over from the PVE 6 installation on the same host which experienced
>>>     a SSD
>>>     >     failure and thus was re-installed with PVE 7.
>>>     >
>>>     >
>>>     >     Am I missing something? Am I being blind?
>>>     >
>>>     >
>>>     >     Any help is appreciated.
>>>     >
>>>     >
>>>     >     Thanks,
>>>     >
>>>     >             Uwe
>>>     >
>>>     >
>>>     >
>>>     >
>>>     >     _______________________________________________
>>>     >     pve-user mailing list
>>>     >     pve-user@lists.proxmox.com
>>> <mailto:pve-user@lists.proxmox.com> <mailto:pve-user@lists.proxmox.com
>>> <mailto:pve-user@lists.proxmox.com>>
>>>     <mailto:pve-user@lists.proxmox.com
>>> <mailto:pve-user@lists.proxmox.com> <mailto:pve-user@lists.proxmox.com
>>> <mailto:pve-user@lists.proxmox.com>>>
>>>     >     https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>> <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
>>>     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>> <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>>
>>>     >     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>> <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
>>>     <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>> <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>>>
>>>     >
>>>
>>
>>
>> _______________________________________________
>> pve-user mailing list
>> pve-user@lists.proxmox.com <mailto:pve-user@lists.proxmox.com>
>> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>> <https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user>
> 




^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2021-07-21 18:06 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-07-21 15:17 [PVE-User] PVE 7 & LACP Uwe Sauter
2021-07-21 15:45 ` Gilberto Ferreira
2021-07-21 15:49   ` Uwe Sauter
2021-07-21 16:41     ` Gilberto Ferreira
2021-07-21 17:31       ` Uwe Sauter
2021-07-21 17:58         ` alexandre derumier
2021-07-21 18:05           ` Uwe Sauter

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox
Service provided by Proxmox Server Solutions GmbH | Privacy | Legal