Here is the relevant portion of the switch config that the customer has shared with me. The ports are configured to vlan 17 as native (untagged) and to only allow vlan 17 at all. Note that this is not vlan 0.
interface Vethernet2424
description server 1/3, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1287 channel 2424
no shutdown
interface Vethernet2426
description server 1/2, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1286 channel 2426
no shutdown
interface Vethernet2428
description server 1/1, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1285 channel 2428
no shutdown
1. On all 3 MAAS nodes, the physical interface enp6s0 is the sole member of bondm which is in bride broam. The vnet interfaces of VMs show up under broam as well. The physical interfaces are not vlan tagged.
2. The netplan on each machine looks like this (with differing addresses and customer specific nameserver info):
4. The VMs inside all the pods can communicate with each other.
root@fce:~/fibernet-fcb# juju machines -m controller
Machine State DNS Inst id Series AZ Message
0 started 10.17.101.23 p6aaff bionic default Deployed
1 started 10.17.101.25 84gxpn bionic zone2 Deployed
2 started 10.17.101.24 bqfy3m bionic zone3 Deployed
* Canonical Livepatch is available for installation.
- Reduce system reboots and improve kernel security. Activate at: https://ubuntu.com/livepatch
33 packages can be updated.
0 updates are security updates.
Last login: Fri Nov 16 20:17:20 2018 from 10.17.101.10
ubuntu@juju-1:~$ ping 10.17.101.25 -c 1
PING 10.17.101.25 (10.17.101.25) 56(84) bytes of data.
64 bytes from 10.17.101.25: icmp_seq=1 ttl=64 time=0.534 ms
--- 10.17.101.25 ping statistics ---
1 packets transmitted, 1 received, 0% packet loss, time 0ms
rtt min/avg/max/mdev = 0.534/0.534/0.534/0.000 ms
ubuntu@juju-1:~$ ping 10.17.101.24 -c 1
PING 10.17.101.24 (10.17.101.24) 56(84) bytes of data.
64 bytes from 10.17.101.24: icmp_seq=1 ttl=64 time=0.653 ms
1. Is STP enabled? I don't believe so but I can't find the response from the customer stating so. I will confirm. I don't think this would be an issue, however, because I can see the DHCP responses coming in a timely manner on the virtual interface when I tcpdump.
2. Is STP or a long forward-delay configured on the bridge?
vernhart@infra2:~$ brctl show
bridge name bridge id STP enabled interfaces
broam 8000.fee6cd1cc06b no bondm
vernhart@infra2:~$ brctl showstp broam
broam
bridge id 8000.fee6cd1cc06b
designated root 8000.fee6cd1cc06b
root port 0 path cost 0
max age 20.00 bridge max age 20.00
hello time 2.00 bridge hello time 2.00
forward delay 15.00 bridge forward delay 15.00
ageing time 300.00
hello timer 0.00 tcn timer 0.00
topology change timer 0.00 gc timer 136.28
flags
bondm (1)
port id 8001 state forwarding
designated root 8000.fee6cd1cc06b path cost 100
designated bridge 8000.fee6cd1cc06b message age timer 0.00
designated port 8001 forward delay timer 0.00
designated cost 0 hold timer 0.00
flags
I agree this may not be a MAAS bug specifically but I'm not sure where else to seek assistance.
You say MAAS doesn't support PXE booting over VLANs but vlan 0 is special: https:/ /en.wikipedia. org/wiki/ IEEE_802. 1Q#Frame_ format
"The reserved value 0x000 indicates that the frame does not carry a VLAN ID"
Here is the relevant portion of the switch config that the customer has shared with me. The ports are configured to vlan 17 as native (untagged) and to only allow vlan 17 at all. Note that this is not vlan 0.
interface Vethernet2424
description server 1/3, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1287 channel 2424
no shutdown
interface Vethernet2426
description server 1/2, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1286 channel 2426
no shutdown
interface Vethernet2428
description server 1/1, VNIC eth0
switchport mode trunk
no lldp transmit
no lldp receive
no pinning server sticky
pinning server pinning-failure link-down
switchport trunk native vlan 17
switchport trunk allowed vlan 17
bind interface port-channel1285 channel 2428
no shutdown
1. On all 3 MAAS nodes, the physical interface enp6s0 is the sole member of bondm which is in bride broam. The vnet interfaces of VMs show up under broam as well. The physical interfaces are not vlan tagged.
2. The netplan on each machine looks like this (with differing addresses and customer specific nameserver info):
network:
dhcp4: false
interfaces: [ enp6s0 ]
parameters:
mode: active-backup
primary: enp6s0
addresses: [ 10.17.101.10/22 ]
gateway4: 10.17.100.1
interfaces: [ bondm ]
nameservers:
addresses: [ 123.123.123.1, 123.123.123.2 ]
search: [ unicloud1. example. net ]
ethernets:
enp6s0:
version: 2
bonds:
bondm:
bridges:
broam:
3. The command-line for dhcpd doesn't show an interface:
vernhart@ infra1: ~$ ps fuax | grep dhcpd dhcp/dhcpd. pid -cf /var/lib/ maas/dhcpd. conf -lf /var/lib/ maas/dhcp/ dhcpd.leases broam infra1: ~$ sudo netstat -nlp | grep dhcp
vernhart 24086 0.0 0.0 13136 1100 pts/8 S+ 20:25 0:00 \_ grep --color=auto dhcpd
dhcpd 8794 0.0 0.0 45964 16976 ? Ss Nov29 0:11 dhcpd -user dhcpd -group dhcpd -f -q -4 -pf /run/maas/
vernhart@
tcp 0 0 10.17.101.10:647 0.0.0.0:* LISTEN 8794/dhcpd
tcp 0 0 0.0.0.0:7911 0.0.0.0:* LISTEN 8794/dhcpd
udp 5120 0 0.0.0.0:67 0.0.0.0:* 8794/dhcpd
udp 0 0 0.0.0.0:7309 0.0.0.0:* 8794/dhcpd
udp6 0 0 :::27481 :::* 8794/dhcpd
raw 0 0 0.0.0.0:1 0.0.0.0:* 7 8794/dhcpd
4. The VMs inside all the pods can communicate with each other.
root@ fce:~/fibernet- fcb# juju machines -m controller
Machine State DNS Inst id Series AZ Message
0 started 10.17.101.23 p6aaff bionic default Deployed
1 started 10.17.101.25 84gxpn bionic zone2 Deployed
2 started 10.17.101.24 bqfy3m bionic zone3 Deployed
root@ fce:~/fibernet- fcb# juju ssh -m controller 0
Welcome to Ubuntu 18.04.1 LTS (GNU/Linux 4.15.0-39-generic x86_64)
* Documentation: https:/ /help.ubuntu. com /landscape. canonical. com /ubuntu. com/advantage
* Management: https:/
* Support: https:/
System information as of Fri Nov 30 22:29:54 UTC 2018
System load: 0.06 Processes: 138
Usage of /: 13.2% of 91.17GB Users logged in: 0
Memory usage: 2% IP address for ens6: 10.17.101.23
Swap usage: 0%
Get cloud support with Ubuntu Advantage Cloud Guest: www.ubuntu. com/business/ services/ cloud
http://
* Canonical Livepatch is available for installation. /ubuntu. com/livepatch
- Reduce system reboots and improve kernel security. Activate at:
https:/
33 packages can be updated.
0 updates are security updates.
Last login: Fri Nov 16 20:17:20 2018 from 10.17.101.10
ubuntu@juju-1:~$ ping 10.17.101.25 -c 1
PING 10.17.101.25 (10.17.101.25) 56(84) bytes of data.
64 bytes from 10.17.101.25: icmp_seq=1 ttl=64 time=0.534 ms
--- 10.17.101.25 ping statistics --- 534/0.534/ 0.000 ms
1 packets transmitted, 1 received, 0% packet loss, time 0ms
rtt min/avg/max/mdev = 0.534/0.
ubuntu@juju-1:~$ ping 10.17.101.24 -c 1
PING 10.17.101.24 (10.17.101.24) 56(84) bytes of data.
64 bytes from 10.17.101.24: icmp_seq=1 ttl=64 time=0.653 ms
--- 10.17.101.24 ping statistics --- 653/0.653/ 0.000 ms
1 packets transmitted, 1 received, 0% packet loss, time 0ms
rtt min/avg/max/mdev = 0.653/0.
ubuntu@juju-1:~$
Your final two asks:
1. Is STP enabled? I don't believe so but I can't find the response from the customer stating so. I will confirm. I don't think this would be an issue, however, because I can see the DHCP responses coming in a timely manner on the virtual interface when I tcpdump.
2. Is STP or a long forward-delay configured on the bridge?
vernhart@ infra2: ~$ brctl show infra2: ~$ brctl showstp broam
bridge name bridge id STP enabled interfaces
broam 8000.fee6cd1cc06b no bondm
vernhart@
broam
bridge id 8000.fee6cd1cc06b
designated root 8000.fee6cd1cc06b
root port 0 path cost 0
max age 20.00 bridge max age 20.00
hello time 2.00 bridge hello time 2.00
forward delay 15.00 bridge forward delay 15.00
ageing time 300.00
hello timer 0.00 tcn timer 0.00
topology change timer 0.00 gc timer 136.28
flags
bondm (1)
port id 8001 state forwarding
designated root 8000.fee6cd1cc06b path cost 100
designated bridge 8000.fee6cd1cc06b message age timer 0.00
designated port 8001 forward delay timer 0.00
designated cost 0 hold timer 0.00
flags