OASIS Mailing List ArchivesView the OASIS mailing list archive below
or browse/search using MarkMail.

 


Help: OASIS Mailing Lists Help | MarkMail Help

virtio-dev message

[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]


Subject: Re: [virtio-dev] Re: [PATCH v3 0/5] Support for datapath switching during live migration




On 01/09/2019 05:39 AM, Michael S. Tsirkin wrote:
On Tue, Jan 08, 2019 at 08:55:35PM -0800, si-wei liu wrote:

On 1/7/2019 6:25 PM, Michael S. Tsirkin wrote:

    On Mon, Jan 07, 2019 at 05:45:22PM -0800, si-wei liu wrote:

        On 01/07/2019 03:32 PM, Michael S. Tsirkin wrote:

            On Mon, Jan 07, 2019 at 05:29:39PM -0500, Venu Busireddy wrote:

                Implement the infrastructure to support datapath switching during live
                migration involving SR-IOV devices.

                1. This patch is based off on the current VIRTIO_NET_F_STANDBY feature
                    bit and MAC address device pairing.

                2. This set of events will be consumed by userspace management software
                    to orchestrate all the hot plug and datapath switching activities.
                    This scheme has the least QEMU modifications while allowing userspace
                    software to build its own intelligence to control the whole process
                    of SR-IOV live migration.

                3. While the hidden device model (viz. coupled device model) is still
                    being explored for automatic hot plugging (QEMU) and automatic datapath
                    switching (host-kernel), this series provides a supplemental set
                    of interfaces if management software wants to drive the SR-IOV live
                    migration on its own. It should not conflict with the hidden device
                    model but just offers simplicity of implementation.
And if guest *does* need the switch because e.g. attaching xdp wants to
resent the card?



                Si-Wei Liu (2):
                   vfio-pci: Add FAILOVER_PRIMARY_CHANGED event to shorten downtime during failover
                   pci: query command extension to check the bus master enabling status of the failover-primary device

                Sridhar Samudrala (1):
                   virtio_net: Add VIRTIO_NET_F_STANDBY feature bit.

                Venu Busireddy (2):
                   virtio_net: Add support for "Data Path Switching" during Live Migration.
                   virtio_net: Add a query command for FAILOVER_STANDBY_CHANGED event.

                ---
                Changes in v3:
                   Fix issues with coding style in patch 3/5.

                Changes in v2:
                   Added a query command for FAILOVER_STANDBY_CHANGED event.
                   Added a query command for FAILOVER_PRIMARY_CHANGED event.

            Hmm it looks like all feedback I sent e.g. here:
            https://patchwork.kernel.org/patch/10721571/
            got ignored.

            To summarize I suggest reworking the series adding a new command along
            the lines of (naming is up to you):

            query-pci-master - this returns status for a device
                               and enables a *single* event after
                               it changes

            and then removing all status data from the event,
            just notify about the change and *only once*.

        Why removing all status data from the event?

    To make sure users do not forget to call query-pci-master to
    re-enable more events.

IMO the FAILOVER_PRIMARY_CHANGED event is on the performance path, it's an
overkill to enforce round trip query for each event in normal situations.

        It does not hurt to keep them
        as the FAILOVER_PRIMARY_CHANGED event in general is of pretty low-frequency.

    A malicious guest can make it as frequent as it wants to.
    OTOH there is no way to limit.

Will throttle the event rate (say, limiting to no more than 1 event per second)
And if guest *does* need the switch because e.g. attaching xdp wants to
resent the card?
Device reset during attaching XDP prog doesn't end up with PCI master status reset as far as I see. Even so (say with FLR or slot reset), the ending state would be reflected (without having to do a query) if state's going through quick transitions. And, suppressing event emission until next query seems to make management stack perform worse in this case? Management stack end up having to query commands for couple of times to see if status is settled.


a way to limit (as opposed to control guest behavior) ? The other similar
events that apply rate limiting don't suppress event emission until the next
query at all.
We have some problematic interfaces already, that's true.

Doing so would just cause more events missing. As stated in the
earlier example, we should give guest NIC a chance to flush queued packets even
if ending state is same between two events.
I haven't seen that requirement. I guess a reset just stops processing
buffers rather than flush. Care to repeat?
If for some reason a VF driver runs into some hardware/firmware fault and keeps resetting just attempts to fix itself, wouldn't it be necessary for failover to use the PV path temporarily? Without proper interleaving all packets can be dropped in the worst case. Severe packet drops deteriorate network performance quite badly, and may even lead to fatal errors if storage device is being hosted over the network (e.g. iSCSI) in particular.


        As can be seen other similar low-frequent QMP events do have data carried
        over.

        As this event relates to datapath switching, there's implication to coalesce
        events as packets might not get a chance to send out as nothing would ever
        happen when going through quick transitions like
        disabled->enabled->disabled. I would allow at least few packets to be sent
        over wire rather than nothing. Who knows how fast management can react and
        consume these events?

        Thanks,
        -Siwei

    OK if it's so important for latency let's include data in the event.
    Please add comments explaining that you must always re-run query
    afterwards to make sure it's stable and re-enable more events.

I can add comments describing why we need to carry data in the event, and apply
rate limiting to events. But I don't follow why it must suppress event until
next query.
Rate limiting is fundamentally broken.
What is the issue with rate limiting? Seems pretty good fit to me, as otherwise mgmt stack still need to check the status periodically with some slow rate if experiencing malicious attack.
Try a stress of resets and there goes your promise of low downtime.
Is this a normal user would do? I guess only malicious guest may do so. With this in mind, what's more important is to defend the attack by imposing rate limiter, and downtime is not even a factor to consider in this context. The point is that I view rate limiting as a way to defend malicious attack, and it's a good trade-off as reset itself doesn't happen very often in normal situations. And we can always find a proper rate that could balance both needs well.

BTW, there's no way to promise latency/downtime either way - all is based on best effort with the current datapatch switching scheme that involves userspace management stack. IMHO ideally the only way with some latency/downtime guarantee is to have guest to kick off datapath switching through a control queue message, which would cause guest to exit and host kernel would then take the turn to handle the MAC filter movement in a synchronized context. Or we should wait Intel to fix their legacy drivers.


Let me try to re-state: state query is fundamentally required
because otherwise e.g. management restarts do not work.
Yes, we've added the query command for such purpose (handle management restarts). There's no argue about this.

And it is much better to force management to run query every event
and every restart than just hope it handles racy corner cases correctly.
There's no way to handle those racy corner cases with what you suggested, I see missing packets when switching promiscuous mode on virtio-net, and I see no chance it can be improved with the current scheme.


If we do include data in the event then there is no real latency cost to
that, since management can take action in response to the event then do
the query asynchronously at leasure.
So let me turn it around and say I don't follow why you have objections
to blocking following events until query.
See my explanations above.

thanks,
-Siwei


Thanks,
-Siwei







            upon event management does query-pci-master
            and acts accordingly.





                  hmp.c                          |   5 +++
                  hw/acpi/pcihp.c                |  27 +++++++++++
                  hw/net/virtio-net.c            |  42 +++++++++++++++++
                  hw/pci/pci.c                   |   5 +++
                  hw/vfio/pci.c                  |  60 +++++++++++++++++++++++++
                  hw/vfio/pci.h                  |   1 +
                  include/hw/pci/pci.h           |   1 +
                  include/hw/virtio/virtio-net.h |   1 +
                  include/net/net.h              |   2 +
                  net/net.c                      |  61 +++++++++++++++++++++++++
                  qapi/misc.json                 |   5 ++-
                  qapi/net.json                  | 100 +++++++++++++++++++++++++++++++++++++++++
                  12 files changed, 309 insertions(+), 1 deletion(-)

            ---------------------------------------------------------------------
            To unsubscribe, e-mail: virtio-dev-unsubscribe@lists.oasis-open.org
            For additional commands, e-mail: virtio-dev-help@lists.oasis-open.org


    ---------------------------------------------------------------------
    To unsubscribe, e-mail: virtio-dev-unsubscribe@lists.oasis-open.org
    For additional commands, e-mail: virtio-dev-help@lists.oasis-open.org



---------------------------------------------------------------------
To unsubscribe, e-mail: virtio-dev-unsubscribe@lists.oasis-open.org
For additional commands, e-mail: virtio-dev-help@lists.oasis-open.org




[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]