[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <7F861DC0615E0C47A872E6F3C5FCDDBD05E0B6B1@BPXM14GP.gisp.nec.co.jp>
Date: Fri, 23 Jan 2015 00:32:00 +0000
From: Hiroshi Shimamoto <h-shimamoto@...jp.nec.com>
To: "Skidmore, Donald C" <donald.c.skidmore@...el.com>,
Bjørn Mork <bjorn@...k.no>
CC: David Laight <David.Laight@...LAB.COM>,
"e1000-devel@...ts.sourceforge.net"
<e1000-devel@...ts.sourceforge.net>,
"netdev@...r.kernel.org" <netdev@...r.kernel.org>,
"Choi, Sy Jong" <sy.jong.choi@...el.com>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
Hayato Momma <h-momma@...jp.nec.com>
Subject: RE: [E1000-devel] [PATCH 1/2] if_link: Add VF multicast promiscuous
mode control
> Subject: RE: [E1000-devel] [PATCH 1/2] if_link: Add VF multicast promiscuous mode control
> >
> > "Skidmore, Donald C" <donald.c.skidmore@...el.com> writes:
> >
> > > My hang up is more related to: without the nob to enable it (off by
> > > default) we are letting one VF dictate policy for all the other VFs
> > > and the PF. If one VF needs to be in promiscuous multicast so is
> > > everyone else. Their stacks now needs to deal with all the extra
> > > multicast packets. As you point out this might not be a direct
> > > concern for isolation in that the VM could have 'chosen' to join any
> > > Multicast group and seen this traffic. My concern over isolation is
> > > one VF has chosen that all the other VM now have to see this multicast
> > > traffic.
> >
> > Apologies if this question is stupid, but I just have to ask about stuff I don't
> > understand...
> >
> > Looking at the proposed implementation, the promiscous multicast flag
> > seems to be a per-VF flag:
> >
> > +int ixgbe_ndo_set_vf_mc_promisc(struct net_device *netdev, int vf, bool
> > +setting) {
> > + struct ixgbe_adapter *adapter = netdev_priv(netdev);
> > + struct ixgbe_hw *hw = &adapter->hw;
> > + u32 vmolr;
> > +
> > + if (vf >= adapter->num_vfs)
> > + return -EINVAL;
> > +
> > + adapter->vfinfo[vf].mc_promisc_enabled = setting;
> > +
> > + vmolr = IXGBE_READ_REG(hw, IXGBE_VMOLR(vf));
> > + if (setting) {
> > + e_info(drv, "VF %u: enabling multicast promiscuous\n", vf);
> > + vmolr |= IXGBE_VMOLR_MPE;
> > + } else {
> > + e_info(drv, "VF %u: disabling multicast promiscuous\n", vf);
> > + vmolr &= ~IXGBE_VMOLR_MPE;
> > + }
> > +
> > + IXGBE_WRITE_REG(hw, IXGBE_VMOLR(vf), vmolr);
> > +
> > + return 0;
> > +}
> > +
> >
> > I haven't read the data sheet, but I took a quick look at the excellent high
> > level driver docs:
> > http://www.intel.com/content/dam/doc/design-guide/82599-sr-iov-driver-
> > companion-guide.pdf
> >
> > It mentions "Multicast Promiscuous Enable" in its "Thoughts for
> > Customization" section:
> >
> > 7.1 Multicast Promiscuous Enable
> >
> > The controller has provisions to allow each VF to be put into Multicast
> > Promiscuous mode. The Intel reference driver does not configure this
> > option .
> >
> > The capability can be enabled/disabled by manipulating the MPE field (bit
> > 28) of the PF VF L2 Control Register (PFVML2FLT – 0x0F000)
> >
> > and showing a section from the data sheet describing the "PF VM L2 Control
> > Register - PFVML2FLT[n] (0x0F000 + 4 * n, n=0...63; RW)"
> >
> > To me it looks like enabling Promiscuos Multicast for a VF won't affect any
> > other VF at all. Is this really not the case?
> >
> >
> >
> > Bjørn
>
> Clearly not a dumb question at all and I'm glad you mentioned that. :) I was going off the assumption, been awhile since
> I read the patch, that the patch was using FCTRL.MPE or MANC.MCST_PASS_L2 which would turn multicast promiscuous on for
> everyone. Since the patch is using PFVML2FLT.MPE this lessens my concern over effect on the entire system.
I believe the patches for this VF multicast promiscuous mode is per VF.
>
> That said I still would prefer having a way to override this behavior on the PF, although I admit my argument is weaker.
> I'm still concerned about a VF changing the behavior of the PF without any way to prevent it. This might be one part
> philosophical (PF sets policy not the VF) but this still could have a noticeable effect on the overall system. If any
> other VFs (or the PF) are receiving MC packets these will have to be replicated which will be a performance hit. When
> we use the MC hash this is limited vs. when anyone is in MC promiscuous every MC packet used by another pool would be
> replicated. I could imagine in some environments (i.e. public clouds) where you don't trust what is running in your VM
> you might what to block this from happening.
I understand your request and I'm thinking to submit the patches
1) Add new mbox API between ixgbe/ixgbevf to turn MC promiscuous on,
and enables it when ixgbevf needs over 30 MC addresses.
2) Add a policy knob to prevent enabling it from the PF.
Does it seem okay?
BTW, I'm bit worried about to use ndo interface for 2) because adding a
new hook makes core code complicated.
Is it really reasonable to do it with ndo?
I haven't find any other suitable method to do it, right now. And using
ndo VF hook looks standard way to control VF functionality.
Then, I think it's the best way to implement this policy in ndo hook.
>
> In some ways it is almost the mirror image of the issue you brought up:
>
> Adding a new hook for this seems over-complicated to me. And it still
> doesn't solve the real problems that
> a) the user has to know about this limit, and
> b) manually configure the feature
>
> My reverse argument might be that if this happens automatically. It might take the VM provider a long time to realize
> performance has taken a hit because some VM asked to join 31 multicast groups and entered MC promiscuous. Then only to
> find that they have no way to block such behavior.
>
> Maybe I wouldn't as concerned if the patch author could provide some performance results to show this won't have as a
> negative effect as I'm afraid it might?
Hm, what kind of test case would you like to have?
The user A who has 30 MC addresses vs the user B who has 31 MC addresses, which
means that MC promiscuous mode, and coming MC packets the user doesn't expect?
thanks,
Hiroshi
Powered by blists - more mailing lists