On 03/16/2017 06:52 PM, Daniel P. Berrange wrote:
On Thu, Mar 16, 2017 at 04:35:36PM +0100, Kevin Wolf wrote:
> Am 16.03.2017 um 16:08 hat Daniel P. Berrange geschrieben:
>> On Thu, Mar 16, 2017 at 06:00:46PM +0300, Denis V. Lunev wrote:
>>> On 03/16/2017 05:45 PM, Daniel P. Berrange wrote:
>>>> On Thu, Mar 16, 2017 at 05:08:57PM +0300, Denis V. Lunev wrote:
>>>>> Hello, All!
>>>>>
>>>>> There is a problem in the current libvirt implementation. domain.xml
>>>>> allows to specify only basic set of options, especially in the case
>>>>> of QEMU, when there are really a lot of tweaks in format drivers.
>>>>> Most likely these options will never be supported in a good way
>>>>> in libvirt as recognizable entities.
>>>>>
>>>>> Right now in order to debug libvirt QEMU VM in production I am using
>>>>> very strange approach:
>>>>> - disk section of domain XML is removed
>>>>> - exact command line options to start the disk are specified at the
end
>>>>> of domain.xml whithin <qemu:commandline> as described by
Stefan
>>>>>
>>>>>
http://blog.vmsplice.net/2011/04/how-to-pass-qemu-command-line-options.html
>>>>>
>>>>> The problem is that when debug is finished and viable combinations
of
>>>>> options is found I can not drop VM in such state in the production.
This
>>>>> is the pain and problem. For example, I have spend 3 days with the
>>>>> VM of one customer which blames us for slow IO in the guest. I have
>>>>> found very good combination of non-standard options which increases
>>>>> disk performance 5 times (not 5%). Currently I can not put this
combination
>>>>> in the production as libvirt does not see the disk.
>>>>>
>>>>> I propose to do very simple thing, may be I am not the first one
here,
>>>>> but it would be nice to allow to pass arbitrary option to the QEMU
>>>>> command line. This could be done in a very generic way if we will
>>>>> allow to specify additional options inside <driver> section
like this:
>>>>>
>>>>> <disk type='file' device='disk'>
>>>>> <driver name='qemu' type='qcow2'
cache='none' io='native'
>>>>> iothread='1'>
>>>>> <option name='l2-cache-size'
value='64M/>
>>>>> <option name='cache-clean-interval'
value='32'/>
>>>>> </driver>
>>>>> <source
file='/var/lib/libvirt/images/rhel7.qcow2'/>
>>>>> <target dev='sda' bus='scsi'/>
>>>>> <address type='drive' controller='0'
bus='0' target='0' unit='0'/>
>>>>> </disk>
>>>>>
>>>>> and so on. The meaning (at least for QEMU) is quite simple -
>>>>> these options will just be added to the end of the -drive command
>>>>> line. The meaning for other drivers should be the same and I
>>>>> think that there are ways to pass generic options in them.
>>>> It is a general policy that we do *not* do generic option passthrough
>>>> in this kind of manner. We always want to represent concepts explicitly
>>>> with named attributes, so that if 2 hypervisors support the same concept
>>>> we can map it the same way in the XML
>>> OK. How could I change L2 cache size for QCOW2 image?
>>>
>>> For 1 Tb disk, fragmented in guest, the performance loss is
>>> around 10 times. 10 TIMES. 1000%. The customer could not
>>> wait until proper fix in the next QEMU release especially
>>> if we are able to provide the kludge specifically for him.
>> We can explicitly allow L2 cache size set in the XML but that
>> is a pretty poor solution to the problem IMHO, as the mgmt
>> application has no apriori knowledge of whether a particular
>> cache size is going to be right for a particular QCow2 image.
>>
>> For a sustainable solution, IMHO this really needs to be fixed
>> in QEMU so it has either a more appropriate default, or if a
>> single default is not possible, have QEMU auto-tune its cache
>> size dynamically to suit the characteristics of the qcow2 image.
> A tradeoff between memory usage and performance is policy, and setting
> policy is the management layer's job, no qemu's. We can try to provide
> good defaults, but they are meant for manual users of qemu. libvirt is
> expected to configure everything exactly as it wants it instead of
> relying on defaults.
The question though is how is an app supposed to figure out what the
optimal setting for cache size is ? It seems to require knowledge
of the level of disk fragmentation and guest I/O patterns, neither
of which are things we can know upfront. Which means any atttempt to
set cache size is little more than ill-informed guesswork
Regards,
Daniel
Funny thing that this information could come from the outside world,
f.e. from the SLA which is dependent from the amount of money the
end-user is paying to the hosting provider.
Den