[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]
Subject: Re: [virtio-dev] [PATCH v4] content: Introduce VIRTIO_NET_F_STANDBY feature
On 11/29/2018 1:17 PM, Michael S. Tsirkin wrote:
Then still you need a new netlink API - effectively it alters the running state of macvtap as it steals certain filters out from the NIC that affects the datapath of macvtap. I assume we talk about some kernel mechanism to do automatic datapath switching without involving userspace management stack/orchestration software. In the kernel's (net core's) view that also needs some weak binding/coordination between the VF and the macvtap for which MAC filter needs to be activated. Still this senses to me a new API rather than tweaking the current and long-existing default behavior and making it work transparently just for this case. Otherwise, without introducing a new API, how does the userspace infer that the running kernel supports this new behavior.On Thu, Nov 29, 2018 at 12:14:46PM -0800, si-wei liu wrote:On 11/28/2018 5:15 PM, Michael S. Tsirkin wrote:On Wed, Nov 28, 2018 at 12:28:42PM -0800, si-wei liu wrote:On 11/28/2018 12:06 PM, Michael S. Tsirkin wrote:On Wed, Nov 28, 2018 at 10:39:55AM -0800, Samudrala, Sridhar wrote:On 11/28/2018 9:35 AM, Michael S. Tsirkin wrote:On Wed, Nov 28, 2018 at 09:31:32AM -0800, Samudrala, Sridhar wrote:On 11/28/2018 9:08 AM, Michael S. Tsirkin wrote:On Mon, Nov 26, 2018 at 12:22:56PM -0800, Samudrala, Sridhar wrote:Update: I have just set the vf mac's address to 0 (ip link set ens2f0 vf 1 mac 00:00:00:00:00:00) after unplugging it (the primary device) and the pings started working again on the failover interface. So it seems like the frames were arriving to the vf on the host.Yes. When the VF is unplugged, you need to reset the VFs MAC so that the packets with VMs MAC start flowing via VF, bridge and the virtio interface. Have you looked at this documentation that shows a sample script to initiate live migration? https://www.kernel.org/doc/html/latest/networking/net_failover.html -SridharInteresting I didn't notice it does this. So in fact just defining VF mac will immediately divert packets to the VF? Given guest driver did not initialize VF yet won't a bunch of packets be dropped?There is typo in my stmt above (VF->PF) When the VF is unplugged, you need to reset the VFs MAC so that the packets with VMs MAC start flowing via PF, bridge and the virtio interface. When the VF is plugged in, ideally the MAC filter for the VF should be added to the HW once the guest driver comes up and can receive packets. Currently with intel drivers, the filter gets added to HW as soon as the host admin sets the VFs MAC via ndo_set_vf_mac() api. So potentially there could be packet drops until the VF driver comes up in the VM.Can this be fixed in the intel drivers?I just checked and it looks like this seems to have been addressed in the ice 100Gb driver. Will bring this up issue internally to see if we can change this behavior in i40e/ixgbe drivers.Also what happens if the mac is programmed both in PF (e.g. with macvtap) and VF? Ideally VF will take precedence.I'm seriously doubtful that legacy Intel NIC hardware can do that instead of mucking around with software workaround in the PF driver. Actually, the same applies to other NIC vendors when hardware sees duplicate filters. There's no such control of precedence on one over the other. -SiweiWell removing a MAC from the PF filter when we are adding it to the VF filter should always be possible. Need to keep it in a separate list and re-add it when removing the MAC from VF filter. This can be handled in the net core, no need for driver specific hacks.So that is what I ever said - essentially what you need is a netdev API, rather than to add dirty hacks on each driver. That is fine, but how would you implement it? Note there's no equivalent driver level .ndo API to "move" filters, and all existing .ndo APIs manipulate at the MAC address level as opposed to filters. Are you going to convince netdev this is the right thing to do and we should add such API to the net core and each individual driver?There's no need for a new API IMO. You drop it from list of uc macs, then call .ndo_set_rx_mode.
Well, I specifically called it out one year ago as this work was started that macvtap is what we look into (we don't care about bridge with promiscuous enabled) and the answer I got at the point was that the current model would work well for macvtap too (which I've been very doubtful from the very beginning). Eventually turns out this is not true and it looks like this is slowly converging to what Hyper-V netvsc already supported quite a few years if not a decade ago, sighs...This can be done without changing existing drivers.Still, let's prioritize things correctly. IMHO it's fine if we initially assume promisc mode on the PF. macvlan has this mode too after all.I'm not sure what promisc mode you talked about. As far as I understand it for macvlan/macvtap the NIC is only put into promisc mode when running out of MAC filter entries. Before that all MAC addresses will be added to the NIC as unicast filters. In addition, people prefer macvlan/macvtap for adding isolation in a multi-tenant cloud as well as avoiding performance penalty due to noisy neighbors. I'd rather to hear that claim to be that the current MAC-based pairing scheme doesn't work well with macvtap and only works with bridged setup which has promisc enabled. That would be more helpful for people to understand the situation better. Thanks, -SiweiAs a first step that's fine.
Suppose this behavior is fixable in legacy Intel NIC, you would still need to evacuate the filter programmed by macvtap previously when VF's filter gets activated (typically when VF's netdev is netif_running() in a Linux guest). That's what we and NetVSC call as "datapath switching", and where this could be handled (driver, net core, or userspace) is the core for the architectural design that I spent much time on.Still this assumes just creating a VF doesn't yet program the on-card filter to cause packet drops.
Having said it, I don't expect or would desperately wait on one vendor to fix a legacy driver which wasn't quite motivated, then no work would be done on that. If you'd go the way, please make sure Intel could change their driver first.
Where do you envision the new attribute could be at? Supposedly it'd be exposed by the kernel, which constitutes a new API or API changes.Let's assume drivers are fixed to do that. How does userspace know that's the case? We might need some kind of attribute so userspace can detect it.
Question is how does userspace know driver isn't broken in this respect? Let's add a "vf failover" flag somewhere so this can be probed?