
On Fri, Dec 06, 2019 at 11:20:38PM +0800, Alex Williamson wrote:
On Fri, 6 Dec 2019 01:04:07 -0500 Yan Zhao <yan.y.zhao@intel.com> wrote:
On Fri, Dec 06, 2019 at 07:55:30AM +0800, Alex Williamson wrote:
On Wed, 4 Dec 2019 22:26:50 -0500 Yan Zhao <yan.y.zhao@intel.com> wrote:
Dynamic trap bar info region is a channel for QEMU and vendor driver to communicate dynamic trap info. It is of type VFIO_REGION_TYPE_DYNAMIC_TRAP_BAR_INFO and subtype VFIO_REGION_SUBTYPE_DYNAMIC_TRAP_BAR_INFO.
This region has two fields: dt_fd and trap. When QEMU detects a device regions of this type, it will create an eventfd and write its eventfd id to dt_fd field. When vendor drivre signals this eventfd, QEMU reads trap field of this info region. - If trap is true, QEMU would search the device's PCI BAR regions and disable all the sparse mmaped subregions (if the sparse mmaped subregion is disablable). - If trap is false, QEMU would re-enable those subregions.
A typical usage is 1. vendor driver first cuts its bar 0 into several sections, all in a sparse mmap array. So initally, all its bar 0 are passthroughed. 2. vendor driver specifys part of bar 0 sections to be disablable. 3. on migration starts, vendor driver signals dt_fd and set trap to true to notify QEMU disabling the bar 0 sections of disablable flags on. 4. QEMU disables those bar 0 section and hence let vendor driver be able to trap access of bar 0 registers and make dirty page tracking possible. 5. on migration failure, vendor driver signals dt_fd to QEMU again. QEMU reads trap field of this info region which is false and QEMU re-passthrough the whole bar 0 region.
Vendor driver specifies whether it supports dynamic-trap-bar-info region through cap VFIO_PCI_DEVICE_CAP_DYNAMIC_TRAP_BAR in vfio_pci_mediate_ops->open().
If vfio-pci detects this cap, it will create a default dynamic_trap_bar_info region on behalf of vendor driver with region len=0 and region->ops=null. Vvendor driver should override this region's len, flags, rw, mmap in its vfio_pci_mediate_ops.
TBH, I don't like this interface at all. Userspace doesn't pass data to the kernel via INFO ioctls. We have a SET_IRQS ioctl for configuring user signaling with eventfds. I think we only need to define an IRQ type that tells the user to re-evaluate the sparse mmap information for a region. The user would enumerate the device IRQs via GET_IRQ_INFO, find one of this type where the IRQ info would also indicate which region(s) should be re-evaluated on signaling. The user would enable that signaling via SET_IRQS and simply re-evaluate the ok. I'll try to switch to this way. Thanks for this suggestion.
sparse mmap capability for the associated regions when signaled.
Do you like the "disablable" flag of sparse mmap ? I think it's a lightweight way for user to switch mmap state of a whole region, otherwise going through a complete flow of GET_REGION_INFO and re-setup region might be too heavy.
No, I don't like the disable-able flag. At what frequency do we expect regions to change? It seems like we'd only change when switching into and out of the _SAVING state, which is rare. It seems easy for userspace, at least QEMU, to drop the entire mmap configuration and ok. I'll try this way.
re-read it. Another concern here is how do we synchronize the event? Are we assuming that this event would occur when a user switch to _SAVING mode on the device? That operation is synchronous, the device must be in saving mode after the write to device state completes, but it seems like this might be trying to add an asynchronous dependency. Will the write to device_state only complete once the user handles the eventfd? How would the kernel know when the mmap re-evaluation is complete. It seems like there are gaps here that the vendor driver could miss traps required for migration because the user hasn't completed the mmap transition yet. Thanks,
Alex
yes, this asynchronous event notification will cause vendor driver miss traps. But it's supposed to be of very short period time. That's also a reason for us to wish the re-evaluation to be lightweight. E.g. if it's able to be finished before the first iterate, it's still safe. But I agree, the timing is not guaranteed, and so it's best for kernel to wait for mmap re-evaluation to complete. migration_thread |->qemu_savevm_state_setup | |->ram_save_setup | | |->migration_bitmap_sync | | |->kvm_log_sync | | |->vfio_log_sync | | | |->vfio_save_setup | |->set_device_state(_SAVING) | |->qemu_savevm_state_pending | |->ram_save_pending | | |->migration_bitmap_sync | | |->kvm_log_sync | | |->vfio_log_sync | |->vfio_save_pending | |->qemu_savevm_state_iterate | |->ram_save_iterate //send pages | |->vfio_save_iterate ... Actually, we previously let qemu trigger the re-evaluation when migration starts. And now the reason for we to wish kernel to trigger the mmap re-evaluation is that there're other two possible use cases: (1) keep passing through devices when migration starts and track dirty pages using hardware IOMMU. Then when migration is about to complete, stop the device and start trap PCI BARs for software emulation. (we made some changes to let device stop ahead of vcpu ) (2) performance optimization. There's an example in GVT (mdev case): PCI BARs are passed through on vGPU initialization and are mmaped to a host dummy buffer. Then after initialization done, start trap of PCI BARs of vGPUs and start normal host mediation. The initial pass-through can save 1000000 times of mmio trap. Thanks Yan