[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]
Subject: Re: [virtio-dev] Re: [PATCH RFC] virtio-net: enable configurable tx queue size
On 05/23/2017 02:24 PM, Jason Wang wrote:
On 2017年05月23日 13:15, Wei Wang wrote:On 05/23/2017 10:04 AM, Jason Wang wrote:On 2017年05月22日 19:52, Wei Wang wrote:On 05/20/2017 04:42 AM, Michael S. Tsirkin wrote:On Fri, May 19, 2017 at 10:32:19AM +0800, Wei Wang wrote:This should be a generic ring feature, not one specific to virtio net.This patch enables the virtio-net tx queue size to be configurable between 256 (the default queue size) and 1024 by the user. The queue size specified by the user should be power of 2. Setting the tx queue size to be 1024 requires the guest driver to support the VIRTIO_NET_F_MAX_CHAIN_SIZE feature.OK. How about making two more changes below: 1) make the default tx queue size = 1024 (instead of 256).As has been pointed out, you need compat the default value too in this case.The driver gets the size info from the device, then would it cause any compatibility issue if we change the default ring size to 1024 in thevhost case? In other words, is there any software (i.e. any virtio-net driver)functions based on the assumption of 256 queue size?I don't know. But is it safe e.g we migrate from 1024 to an older qemu with 256 as its queue size?
Yes, I think it is safe, because the default queue size is used when the device is being
set up (e.g. feature negotiation).During migration (the device has already been running), the destination machine will load the device state based on the the queue size that is being used (i.e. vring.num).
The default value is not used any more after the setup phase.
For live migration, the queue size that is being used will also be transferredto the destination.We can reduce the size (to 256) if the MAX_CHAIN_SIZE feature is not supported by the guest. In this way, people who apply the QEMU patch can directly use the largest queue size(1024) without adding the booting command line. 2) The vhost backend does not use writev, so I think when the vhost backed is used, using 1024 queue size should not depend on the MAX_CHAIN_SIZE feature.But do we need to consider even larger queue size now?Need Michael's feedback on this. Meanwhile, I'll get the next version of code ready and check if larger queue size would cause any corner case.The problem is, do we really need a new config filed for this? Or just introduce a flag which means "I support up to 1024 sgs" is sufficient?
For now, it also works without the new config field, max_chain_size, But I would prefer to keep the new config field, because: Without that, the driver will work on an assumed value, 1023. If the future, QEMU needs to change it to 1022, then how can the new QEMU tell the old driver, which supports the MAX_CHAIN_SIZE feature but works with the old hardcode value 1023? So, I think using that config value would be good for future updates. Best, Wei
[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]