From: duluxoz <duluxoz@gmail.com>
To: Stefan Hanreich <s.hanreich@proxmox.com>,
Proxmox VE user list <pve-user@lists.proxmox.com>
Subject: Re: [PVE-User] Problem With Bond & VLANs - Help Please
Date: Fri, 16 Aug 2024 20:42:46 +1000 [thread overview]
Message-ID: <534f2488-4441-43eb-9767-5e20b531f6b3@gmail.com> (raw)
In-Reply-To: <822bb310-e82f-4690-bf83-a73cbe957959@proxmox.com>
Hi Stephan,
My apologises, I should have been more precise.
What doesn't work? Most of the ifaces are down (won't come up
automatically as I expect (not even NIC3)), and so I have no
connectivity to the LAN, let alone the rest of the outside world.
Yes, each VLAN should have its own gateway - each VLAN is its own
subnet, of course.
Results of `ip r`:
~~~
default via 10.0.200.1 dev vmbr0 proto kernal onlink linkdown
10.0.100.0/24 dev bond0.100 proto kernal scope link src 10.0.100.0 linkdown
10.0.200.0/24 dev bond0.200 proto kernal scope link src 10.0.200.0 linkdown
10.0.200.0/24 dev vmbr0 proto kernal scope link src 10.0.200.100 linkdown
~~~
Results of `ip a`:
~~~
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host noprefixroute
valid_lft forever preferred_lft forever
2: eno0: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
default qlen 1000
link/ether 00:1b:21:e4:a6:f4 brd ff:ff:ff:ff:ff:ff
3: eno1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
default qlen 1000
link/ether 00:1b:21:e4:a6:f5 brd ff:ff:ff:ff:ff:ff
4: eno2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
default qlen 1000
link/ether 00:1b:21:e4:a6:f6 brd ff:ff:ff:ff:ff:ff
5: bond0: <NO-CARRIER,BROADCAST,MULTICAST,MASTER,UP> mtu 1500 qdisc
noqueue master vmbr0 state DOWN group default qlen 1000
link/ether 4a:3a:67:59:ac:d3 brd ff:ff:ff:ff:ff:ff
6: bond0.100@bond0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
noqueue state LOWERLAYERDOWN group default qlen 1000
link/ether 4a:3a:67:59:ac:d3 brd ff:ff:ff:ff:ff:ff
inet 10.0.100.0/24 scope global bond0.100
valid_lft forever preferred_lft forever
7: bond0.200@bond0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
noqueue state LOWERLAYERDOWN group default qlen 1000
link/ether 4a:3a:67:59:ac:d3 brd ff:ff:ff:ff:ff:ff
inet 10.0.200.0/24 scope global bond0.200
valid_lft forever preferred_lft forever
8: vmbr0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc noqueue
state DOWN group default qlen 1000
link/ether 4a:3a:67:59:ac:d3 brd ff:ff:ff:ff:ff:ff
inet 10.0.200.100/24 scope global vmbr0
valid_lft forever preferred_lft forever
~~~
Thanks for taking a look
Cheers
dulux-oz
On 16/8/24 19:53, Stefan Hanreich wrote:
>
> On 8/16/24 09:36, duluxoz wrote:
>> Hi All,
>>
>> Disclaimer: I'm coming from an EL background - this is my first venture
>> into Debian-world :-)
>>
>> So I'm having an issue getting the NICs, Bond, and VLANs correctly
>> configured on a new Proxmox Node (Old oVirt Node). This worked on the
>> old oVirt config (abit a different set of config files/statements).
>>
>> What I'm trying to achieve:
>>
>> * Proxmox Node IP Address: 10.0.200.100/24, Tag:VLAN 200
>> * Gateway: 10.0.200.1
>> * Bond: NIC1 (eno0) & NIC2 (eno1), 802.3ad
>> * VLAN bond0.100: 10.0.100.0/24, Gateway 10.0.100.1
>> * VLAN bond0.200: 10.0.200.0/24, Gateway 10.0.200.1
>> * NIC3 (eno2): 10.0.300.100/24 - not really relevant, as its not part
>> of the Bond, but I've included it to be thorough
> What *exactly* doesn't work?
> Does the configuration not apply? Do you not get any connectivity with
> internet / specific networks?
>
>
> First thing that springs to mind is that you cannot configure two
> default gateways. There can only be one default gateway. You can
> configure different gateways for different subnets / interfaces. Or you
> can configure different routing tables for different processes.
>
> Your current configuration specifies three gateways. I assume you want
> to use different gateways for different subnets?
>
>
> How does the output of the following commands look like?
>
> ip a
> ip r
_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
next prev parent reply other threads:[~2024-08-16 10:43 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-08-16 7:36 duluxoz
2024-08-16 8:05 ` Christian Kivalo
2024-08-16 9:43 ` duluxoz
2024-08-16 9:53 ` Stefan Hanreich
2024-08-16 10:42 ` duluxoz [this message]
2024-08-16 11:26 ` Gilberto Ferreira
2024-08-16 11:26 ` proxmox
2024-08-16 11:32 ` proxmox
2024-08-18 11:14 duluxoz
2024-08-18 14:07 ` Gilberto Ferreira
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=534f2488-4441-43eb-9767-5e20b531f6b3@gmail.com \
--to=duluxoz@gmail.com \
--cc=matthew@peregrineit.net \
--cc=pve-user@lists.proxmox.com \
--cc=s.hanreich@proxmox.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.