Hi Daniel,
I thought migration might be the reason, but I'm still not seeing the
behavior you describe with regards to pausing. I saw the following behavior:
1. Created VM on node 1
2. Started VM on node 1
3. Migrated VM to node 2, node 1 is now shutdown, node 2 is running
4. I paused node 2
5. I started node 1, no error
6. Paused node 1
7. Unpaused node 2, no err
I thought maybe the original VM had to be paused first, so I tried that
as well:
1. Created VM on node 1
2. Started VM on node 1
3. Migrated to node 2, node 1 is now shutdown, node 2 is running
4. I shutdown node 2 instead of pausing
5. I started node 1
6. I paused node 1
7. Started node 2
8. Paused node 2
9. Started node 1
So sanlock is preventing both from running concurrently, but it seems to
contradict your statement:
"Even if you now pause the VM on node 2, and try to resume node 1,
sanlock will still prevent node 1 from running again. It tracks a lock
"version" number. The key is that once the original VM is paused, if any
other VM runs on that disk, the orignal VM is never allowed to be
unpaused again."
Am I mistaken?
Thanks,
Michael
On 1/31/2013 9:44 AM, Daniel P. Berrange wrote:
On Thu, Jan 31, 2013 at 09:34:47AM -0800, Michael Rodrigues wrote:
> Hello,
>
> I'm using libvirt and sanlock on qemu-kvm guests. Each guest has
> it's own Logical Volume for it's root filesystem. Sanlock is
> configured and working and prevents me from starting the same VM
> twice on multiple nodes and corrupting it's root filesystem. Each
> VM's domain XML resides on 2 servers that share the LVM volume group
> over fiber channel.
>
> In testing, I noticed that if I pause a VM on node 1, the sanlock
> lock is relinquished, and I am able to start the same VM, using the
> same root filesystem, on node 2. I get a lock error when unpausing
> node 1's VM if node 2's copy is still running, but by this point,
> the disk may already be corrupted.
The disk isn't corrupted - preventing node 1 VM from un-pausing
is explicitly what is preventing the disk from becoming corrupted.
Even if you now pause the VM on node 2, and try to resume node 1,
sanlock will still prevent node 1 from running again. It tracks
a lock "version" number. The key is that once the original VM
is paused, if any other VM runs on that disk, the orignal VM is
never allowed to be unpaused again.
> Is it necessary that paused VMs don't get to keep their locks? Is
> there a way to configure sanlock to lock when a VM is paused?
No, it isn't something that's configurable - this behaviour is
required in order todo migration for example.
Regards,
Daniel
--
Michael Rodrigues
Interim Help Desk Manager
Gevirtz Graduate School of Education
Education Building 4203
(805) 893-8031
help(a)education.ucsb.edu