Hello guys,
maybe you can help me with following issue. I have created a little cloud with a host and two worker nodes using opennebula. The setup went successfully until now, I am able to create VM's and move them via normal and live migration.
Another (possibly) important information is that I configured my virtual bridge on both worker nodes like this:

 auto br0
 iface br0 inet static
         address 192.168.0.[2|3]
         netmask 255.255.255.0
         network 192.168.0.0
         broadcast 192.168.0.255
         #gateway 192.168.0.1
         bridge_ports eth0
         bridge_stp on
         bridge_maxwait 0

The command "brctl show" gives me following things back:

bridge name    bridge id            STP enabled    interfaces
br0
            8000.003005c34278    yes            eth0
                                                   vnet0 (<- only appears on node with running VM)

virbr0         8000.000000000000    yes

According to the libvirt wiki this setting is good as is. However, the issue I'm having is that when I create a VM and assign a static IP to it, which looks like e.g. 192.168.0.5,
I firstly am able to ping this VM from both
worker nodes, and also when I perform a live migration the ping stops for a few seconds (until the nodes realize the new route to this VM) and then
starts pinging normally again.

However, when I perform a normal migration the ping doesn't recover anymore, but answers repeatedly with: Destination Host Unreachable

Do you know what could be the problem? Where is the difference between a normal and live migration and how can the ping after live migrating still work, but after a normal migration not?

Thanks a lot!
Regards, Adnan