[dpdk-dev] [PATCH] net/bonding: propagate promiscous mode in mode 4

Matan Azrad matan at mellanox.com
Mon Sep 17 08:29:02 CEST 2018


Hi Chas

From:  Chas Williams
> On Thu, Sep 13, 2018 at 11:40 AM Matan Azrad <matan at mellanox.com>
> wrote:
> >
> > Hi Chas
> >
> > From:  Chas Williams
> > > On Wed, Sep 12, 2018 at 1:56 AM Matan Azrad <matan at mellanox.com>
> > > wrote:
> > > >
> > > > Hi Chas
> > > >
> > > > From:  Chas Williams
> > > > > On Mon, Aug 6, 2018 at 3:35 PM Matan Azrad
> <matan at mellanox.com>
> > > > > wrote:
> > > > > >
> > > > > >
> > > > > > Hi Chas
> > > > > >
> > > > > > From: Chas Williams
> > > > > > >On Mon, Aug 6, 2018 at 1:46 PM Matan Azrad
> > > > > <mailto:matan at mellanox.com> wrote:
> > > > > > >Hi Chas
> > > > > > >
> > > > > > >From: Chas Williams
> > > > > > >>On Fri, Aug 3, 2018 at 1:47 AM Matan Azrad
> > > > > <mailto:mailto:matan at mellanox.com> wrote:
> > > > > > >>Hi Chas
> > > > > > >>
> > > > > > >> From: Chas Williams
> > > > > > >> [mailto:mailto:mailto:mailto:3chas3 at gmail.com]
> > > > > > >> On Thu, Aug 2, 2018 at 1:33
> > > > > > >>> PM Matan Azrad <mailto:mailto:matan at mellanox.com> wrote:
> > > > > > >>> >
> > > > > > >>> > > I suggest to do it like next, To add one more
> > > > > > >>> > > parameter for LACP which means how to configure the
> > > > > > >>> > LACP MC group - lacp_mc_grp_conf:
> > > > > > >>> > > 1. rte_flow.
> > > > > > >>> > > 2. flow director.
> > > > > > >>> > > 3. add_mac.
> > > > > > >>> > > 3. set_mc_add_list
> > > > > > >>> > > 4. allmulti
> > > > > > >>> > > 5. promiscuous
> > > > > > >>> > > Maybe more... or less :)
> > > > > > >>> > >
> > > > > > >>> > > By this way the user decides how to do it, if it's
> > > > > > >>> > > fail for a slave, the salve
> > > > > > >>> > should be rejected.
> > > > > > >>> > > Conflict with another configuration(for example
> > > > > > >>> > > calling to promiscuous
> > > > > > >>> > disable while running LACP lacp_mc_grp_conf=5) should
> > > > > > >>> > raise an
> > > > > error.
> > > > > > >>> > >
> > > > > > >>> > > What do you think?
> > > > > > >>> > >
> > > > > > >>> >
> > > > > > >>> > Supporting an LACP mc group specific configuration does
> > > > > > >>> > make sense, but I wonder if this could just be handled
> > > > > > >>> > by default during
> > > > > slave add.
> > > > > > >>> >
> > > > > > >>> >
> > > > > > >>> > 1 and 2 are essentially the same hardware filtering
> > > > > > >>> > offload mode, and the other modes are irrelevant if this
> > > > > > >>> > is enabled, it should not be possible to add the slave
> > > > > > >>> > if the bond is configured for this mode, or possible to
> > > > > > >>> > change the bond into this mode if an existing slave doesn't
> support it.
> > > > > > >>>
> > > > > > >>> >
> > > > > > >>> > 3 should be the default expected behavior, but
> > > > > > >>> > rte_eth_bond_slave_add() should fail if the slave being
> > > > > > >>> > added doesn't support either adding the MAC to the slave
> > > > > > >>> > or adding the
> > > > > LACP MC address.
> > > > > > >>> >
> > > > > > >>> > Then the user could try either
> > > > > > >>> > rte_eth_allmulticast_enable() on the bond port and then
> > > > > > >>> > try to add the slave again, which should fail if
> > > > > > >>> > existing slave didn't support allmulticast or the add
> > > > > > >>> > slave would fail again if the slave didn't support
> > > > > > >>> > allmulticast and finally just call
> > > > > > >>> > rte_eth_promiscuous_enable() on the bond and then try to
> > > > > > >>> > re-add the that slave.
> > > > > > >>> >
> > > > > > >>> > but maybe having a explicit configuration parameter
> > > > > > >>> > would be
> > > > > better.
> > > > > > >>>
> > > > > > >>> I don't sure you understand exactly what I’m suggesting
> > > > > > >>> here,
> > > again:
> > > > > > >>> I suggest to add a new parameter to the LACP mode called
> > > > > > >>> lacp_mc_grp_conf(or something else).
> > > > > > >>> So, when the user configures LACP (mode 4) it must to
> > > > > > >>> configure the lacp_mc_grp_conf parameter to one of the
> > > > > > >>> options I
> > > suggested.
> > > > > > >>> This parameter is not per slave means the bond PMD will
> > > > > > >>> use the selected option to configure the LACP MC group for
> > > > > > >>> all the
> > > slave ports.
> > > > > > >>>
> > > > > > >>> If one of the slaves doesn't support the selected option
> > > > > > >>> it should be
> > > > > rejected.
> > > > > > >>> Conflicts should rais an error.
> > > > > > >>>
> > > > > > >>> I agree here.  Yes, if a slave can't manage to subscribe
> > > > > > >>> to the multicast group, an error should be raised.  The
> > > > > > >>> only way for this to happen is that you don't have promisc
> > > > > > >>> support which is the ultimate
> > > > > fallback.
> > > > > > >>
> > > > > > >>> The advantages are:
> > > > > > >>> The user knows which option is better to synchronize with
> > > > > > >>> his
> > > > > application.
> > > > > > >>> The user knows better than the bond PMD what is the slaves
> > > > > capabilities.
> > > > > > >>> All the slaves are configured by the same way - consistent
> traffic.
> > > > > > >>>
> > > > > > >>>
> > > > > > >>> It would be ideal if all the slaves would have the same
> > > > > > >>> features and capabilities.  There wasn't enforced before,
> > > > > > >>> so this would be a new restriction that would be less
> > > > > > >>> flexible than what we currently have.  That doesn't seem
> > > > > > >>> like an
> > > improvement.
> > > > > > >>
> > > > > > >>> The bonding user probably doesn't care which mode is used.
> > > > > > >>> The bonding user just wants bonding to work.  He doesn't
> > > > > > >>> care about
> > > > > the details.   If I am writing
> > > > > > >>> an application with this proposed API, I need to make a
> > > > > > >>> list of adapters and what they support (and keep this up
> > > > > > >>> to date as DPDK
> > > > > evolves).  Ugh.
> > > > > > >>
> > > > > > >>The applications commonly know what are the nics
> > > > > > >>capabilities they
> > > > > work with.
> > > > > > >>
> > > > > > >>I know at least an one big application which really
> > > > > > >>suffering because the bond configures promiscuous in mode 4
> > > > > > >>without the
> > > > > application asking (it's considered there as a bug in dpdk).
> > > > > > >>I think that providing  another option will be better.
> > > > > > >>
> > > > > > >>I think providing another option will be better as well.
> > > > > > >>However we
> > > > > disagree on the option.
> > > > > > >>If the PMD has no other way to subscribe the multicast
> > > > > > >>group, it has to
> > > > > use promiscuous mode.
> > > > > > >
> > > > > > >>Yes, it is true but there are a lot of other and better
> > > > > > >>options,
> > > > > promiscuous is greedy! Should be the last alternative to use.
> > > > > > >
> > > > > > >Unfortunately, it's the only option implemented.
> > > > > >
> > > > > > Yes, I know, I suggest to change it or at least not to make it worst.
> > > > > >
> > > > > > >>Providing a list of options only makes life complicated for
> > > > > > >>the developer and doesn't really make any difference in the end
> results.
> > > > > > >
> > > > > > >>A big different, for example:
> > > > > > >>Let's say the bonding groups 2 devices that support rte_flow.
> > > > > > >>The user don't want neither promiscuous nor all multicast,
> > > > > > >>he just want to get it's mac traffic + LACP MC group
> > > > > > >>traffic,(a realistic use case)  if
> > > > > he has an option to tell to the bond PMD, please use rte_flow
> > > > > to configure the specific LACP MC group it will be great.
> > > > > > >>Think how much work these applications should do in the
> > > > > > >>current
> > > > > behavior.
> > > > > > >
> > > > > > >The bond PMD should already know how to do that itself.
> > > > > >
> > > > > > The bond can do it with a lot of complexity, but again the
> > > > > > user must know
> > > > > what the bond chose to be synchronized.
> > > > > > So, I think it's better that the user will define it because
> > > > > > it is a traffic configuration (the same as promiscuous
> > > > > > configuration - the user configures it)
> > > > > > >  Again, you are forcing more work on the user to ask them to
> > > > > > > select
> > > > > between the methods.
> > > > > >
> > > > > > We can create a default option as now(promiscuous).
> > > > > >
> > > > > > >>  For instance, if the least common denominator between the
> > > > > > >>two PMDs is promiscuous mode,  you are going to be forced to
> > > > > > >>run both in promiscuous mode instead of selecting the best
> > > > > > >>mode for
> > > each PMD.
> > > > > > >
> > > > > > >>In this case promiscuous is better, Using a different
> > > > > > >>configuration is worst and against the bonding PMD
> > > > > principle to get a consistent traffic from the slaves.
> > > > > > >>So, if one uses allmulti and one uses promiscuous the
> > > > > > >>application may get an inconsistent traffic and it may
> > > > > > >>trigger a lot of problems and
> > > > > complications for some applications.
> > > > > > >
> > > > > > >Those applications should already have those problems.
> > > > > > >  I can make the counter
> > > > > > >argument that there are potentially applications relying on
> > > > > > >the broken
> > > > > behavior.
> > > > > >
> > > > > > You right. So adding allmulticast will require changes in
> > > > > > these
> > > applications.
> > > > > >
> > > > > > >We need to ignore those issues and fix this the "right" way.
> > > > > > >The "right" way IMHO is the pass the least amount of traffic
> > > > > > >possible in each
> > > > > case.
> > > > > >
> > > > > > Not in cost of an inconsistency, but looks like we are not agree here.
> > > > > >
> > > > >
> > > > > I have recently run into this issue again with a device that
> > > > > doesn't support promiscuous, but does let me subscribe to the
> > > > > appropriate
> > > multicast groups.
> > > > > At this point, I am leaning toward adding another API call to
> > > > > the bonding API so that the user can provide a callback to setup
> > > > > whatever they want on the slaves.
> > > > > The default setup routine would be enable promiscuous.
> > > > >
> > > > > Comments?
> > > >
> > > > The bonding already allows to the users to do operations directly
> > > > to the
> > > slaves(it exports the port ids - rte_eth_bond_slaves_get), so I
> > > don't understand why do you need a new API.
> > > > The only change you need may be to add parameter to disable the
> > > promiscuous configuration in mode4.
> > >
> > > Changing the API is new API.  We should attempt to not break any of
> > > the existing API.
> > >
> > > As for being able to operate on the slaves, yet, you can.  But the
> > > bonding PMD also controls the slaves as well.  It seems cleaner to
> > > make this explcit, that the bonding driver calls out to the
> > > application to setup the 802.3ad listening when it needs to be done.
> > > If you want to control it a different way, you simply provide a null
> > > routine that does nothing and control it however you like.
> >
> > The issue is that the bonding PMD cannot be synchronized with such like
> callback, it doesn't know what was done by the application.
> 
> Exactly.  That's why you need to be to change the native behavior of the
> bonding PMD.

I don't sure I understand you here,
But I agree that both bonding PMD and the application should be able to know exactly what is the configuration set at any time.
The current behavior is that bonding configures promiscuous in mode4 without any doc\guide which says it to the application user
and it is problematic and need to be changed.
 
> 
> > I don't think we should open a direct application calls to the slaves by an
> API, if the application really need  it, as we said, it has already option to do it
> without a new API while it is really not recommended by the bonding guide.
> 
> This would be the opposite direction.  The slaves would be calling out to the
> application to ask the application to do something.

This "something" must be known by the bonding PMD also - that what I am saying.
In the callback solution you are suggesting now, the application knows but the bonding PMD doesn't.

>  With the existing bonding
> PMD it will always attempt to enable promisc and that isn't desirable in all
> situations.

Agree, it's greedy.

>  Previously, activate_slave() always happened as part of slave add.
> That shouldn't be the case now if the bonding PMD isn't running.  Adding and
> removing slaves while the bonding PMD is started shouldn't probably not be
> done because I am not sure we can ensure that there is race free behavior
> with the 802.3ad rx/tx routines.

I don't sure I understand what you want to say in the above statement.
I didn't talk about synchronization in terms of locks solutions but in terms of "every entity must know what are the slave configurations at any time".


That’s what I suggested in the past, saying it again:
Let the user to decide the way he wants to configure the LACP MC group by the bonding command line:
1. promiscuous (default)
2. allmulti
3. set mc list
4. set mac
....
Using this way both the application user and the bonding PMD know what are the configurations at any time.
(also saves consistency and the last behavior of the bonding)





 




More information about the dev mailing list