OASIS Mailing List ArchivesView the OASIS mailing list archive below
or browse/search using MarkMail.

 


Help: OASIS Mailing Lists Help | MarkMail Help

virtio-dev message

[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]


Subject: Re: [virtio-dev] [PATCH v4] content: Introduce VIRTIO_NET_F_STANDBY feature




On 11/29/2018 1:17 PM, Michael S. Tsirkin wrote:
On Thu, Nov 29, 2018 at 12:14:46PM -0800, si-wei liu wrote:

On 11/28/2018 5:15 PM, Michael S. Tsirkin wrote:
On Wed, Nov 28, 2018 at 12:28:42PM -0800, si-wei liu wrote:
On 11/28/2018 12:06 PM, Michael S. Tsirkin wrote:
On Wed, Nov 28, 2018 at 10:39:55AM -0800, Samudrala, Sridhar wrote:
On 11/28/2018 9:35 AM, Michael S. Tsirkin wrote:
On Wed, Nov 28, 2018 at 09:31:32AM -0800, Samudrala, Sridhar wrote:
On 11/28/2018 9:08 AM, Michael S. Tsirkin wrote:
On Mon, Nov 26, 2018 at 12:22:56PM -0800, Samudrala, Sridhar wrote:
Update:
I have just set the vf mac's address to 0 (ip link set ens2f0 vf 1 mac
00:00:00:00:00:00) after unplugging it (the primary device) and the
pings started working again on the failover interface. So it seems
like the frames were arriving to the vf on the host.


Yes. When the VF is unplugged, you need to reset the VFs MAC so that the packets
with VMs MAC start flowing via VF, bridge and the virtio interface.

Have you looked at this documentation that shows a sample script to initiate live
migration?
https://www.kernel.org/doc/html/latest/networking/net_failover.html

-Sridhar
Interesting I didn't notice it does this. So in fact
just defining VF mac will immediately divert packets
to the VF? Given guest driver did not initialize VF
yet won't a bunch of packets be dropped?
There is typo in my stmt above (VF->PF)
When the VF is unplugged, you need to reset the VFs MAC so that the packets
with VMs MAC start flowing via PF, bridge and the virtio interface.

When the VF is plugged in, ideally the MAC filter for the VF should be added to
the HW once the guest driver comes up and can receive packets. Currently with intel
drivers, the filter gets added to HW as soon as the host admin sets the VFs MAC via
ndo_set_vf_mac() api. So potentially there could be packet drops until the VF driver
comes up in the VM.


Can this be fixed in the intel drivers?
I just checked and it looks like this seems to have been addressed in the
ice 100Gb driver. Will bring this up issue internally to see if we can change this
behavior in i40e/ixgbe drivers.
Also what happens if the mac is programmed both in PF (e.g. with
macvtap) and VF? Ideally VF will take precedence.
I'm seriously doubtful that legacy Intel NIC hardware can do that instead of
mucking around with software workaround in the PF driver. Actually, the same
applies to other NIC vendors when hardware sees duplicate filters. There's
no such control of precedence on one over the other.


-Siwei


Well removing a MAC from the PF filter when we are adding it to the VF
filter should always be possible. Need to keep it in a separate list and
re-add it when removing the MAC from VF filter.  This can be handled in
the net core, no need for driver specific hacks.
So that is what I ever said - essentially what you need is a netdev API,
rather than to add dirty hacks on each driver. That is fine, but how would
you implement it? Note there's no equivalent driver level .ndo API to "move"
filters, and all existing .ndo APIs manipulate at the MAC address level as
opposed to filters. Are you going to convince netdev this is the right thing
to do and we should add such API to the net core and each individual driver?
There's no need for a new API IMO.
You drop it from list of uc macs, then call .ndo_set_rx_mode.
Then still you need a new netlink API - effectively it alters the running state of macvtap as it steals certain filters out from the NIC that affects the datapath of macvtap. I assume we talk about some kernel mechanism to do automatic datapath switching without involving userspace management stack/orchestration software. In the kernel's (net core's) view that also needs some weak binding/coordination between the VF and the macvtap for which MAC filter needs to be activated. Still this senses to me a new API rather than tweaking the current and long-existing default behavior and making it work transparently just for this case. Otherwise, without introducing a new API, how does the userspace infer that the running kernel supports this new behavior.

This can be done without changing existing drivers.

Still, let's prioritize things correctly.  IMHO it's fine if we
initially assume promisc mode on the PF.  macvlan has this mode too
after all.
I'm not sure what promisc mode you talked about. As far as I understand it
for macvlan/macvtap the NIC is only put into promisc mode when running out
of MAC filter entries. Before that all MAC addresses will be added to the
NIC as unicast filters. In addition, people prefer macvlan/macvtap for
adding isolation in a multi-tenant cloud as well as avoiding performance
penalty due to noisy neighbors. I'd rather to hear that claim to be that the
current MAC-based pairing scheme doesn't work well with macvtap and only
works with bridged setup which has promisc enabled. That would be more
helpful for people to understand the situation better.

Thanks,
-Siwei

As a first step that's fine.
Well, I specifically called it out one year ago as this work was started that macvtap is what we look into (we don't care about bridge with promiscuous enabled) and the answer I got at the point was that the current model would work well for macvtap too (which I've been very doubtful from the very beginning). Eventually turns out this is not true and it looks like this is slowly converging to what Hyper-V netvsc already supported quite a few years if not a decade ago, sighs...

  Still this assumes just creating a VF
doesn't yet program the on-card filter to cause packet drops.
Suppose this behavior is fixable in legacy Intel NIC, you would still need to evacuate the filter programmed by macvtap previously when VF's filter gets activated (typically when VF's netdev is netif_running() in a Linux guest). That's what we and NetVSC call as "datapath switching", and where this could be handled (driver, net core, or userspace) is the core for the architectural design that I spent much time on.

Having said it, I don't expect or would desperately wait on one vendor to fix a legacy driver which wasn't quite motivated, then no work would be done on that. If you'd go the way, please make sure Intel could change their driver first.

  Let's
assume drivers are fixed to do that. How does userspace know
that's the case? We might need some kind of attribute so
userspace can detect it.
Where do you envision the new attribute could be at? Supposedly it'd be exposed by the kernel, which constitutes a new API or API changes.


Thanks,
-Siwei

Question is how does userspace know driver isn't broken in this respect?
Let's add a "vf failover" flag somewhere so this can be probed?




[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]