Hi,

Sorry of this is the wrong list, trying to find some help.

I am trying to set up 2 KVM hosts using Debian 9 (connecting to a 3 node Centos 7.4 gluster cluster)  To get the bandwidth I have bonded 2 NICS each end.  However while the network works fine for the Debian VM host the VM guests cannot connect/get an IP using DHCP connecting to br0.  (Same setup on the Centos end works fine however)

On the Debian host if I connect to br1 which is a bridge to a NIC port it works fine.  My problem then seems to be at the bonding layer for libvirt/kvm ?   

Here is my interfaces config, I cannot see the problem.  The VM guest is on the gluster cluster and boots fine just no networking. 

any ideas what I am doing wrong pls?  or a better group/list to go ask in?


auto eno1
iface eno1  inet  manual
#iface eno1  inet static
    bond-master bond0
    #address 192.168.1.62
    #netmask 255.255.255.0
    #gateway 192.168.1.1
    #dns-nameservers 192.168.1.241 192.168.1.104
#    mtu 9000


auto enp1s0f1
iface enp1s0f1 inet manual
        bond-master bond0
#    mtu 9000

auto bond0
iface bond0 inet static
   bond-miimon 100
   bond-mode 6
   bond-updelay 200
   bond-xmit_hash_policy layer3+4
   bond-lacp-rate 1
#   mtu 9000
# address 192.168.1.62
# netmask 255.255.255.0
# gateway 192.168.1.1
# dns-nameservers 192.168.1.241 192.168.1.104
   slaves eno1 enp1s0f1


auto br0
iface br0 inet static
   address 192.168.1.62
   netmask 255.255.255.0
   gateway 192.168.1.1
   dns-nameservers 192.168.1.241 192.168.1.104
   bridge_ports bond0
   #bridge_ports eno1 enp1s0f1
   bridge_stp on
   bridge_fd 0.0
#    mtu 9000
#post-up ifconfig bond0 mtu 9000
#post-up ip link set dev bond0 mtu 9000

#auto bond1
#iface bond1 inet static
#   address 10.100.200.62
#   netmask 255.255.255.0
#   bond_slaves enp3s0f1
#   bond_primary enp3s0f1
#    mtu 9000


auto br1
iface br1 inet static
   address 10.100.200.62
   netmask 255.255.255.0
   gateway 10.100.200.1
   bridge_ports enp3s0f1
   bridge_stp on
   bridge_fd 0.0
root@kvm02:/etc/network#