lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <55198210.2030804@profitbricks.com>
Date:	Mon, 30 Mar 2015 19:04:16 +0200
From:	Michael Wang <yun.wang@...fitbricks.com>
To:	Doug Ledford <dledford@...hat.com>
CC:	Roland Dreier <roland@...nel.org>,
	Sean Hefty <sean.hefty@...el.com>,
	Hal Rosenstock <hal.rosenstock@...il.com>,
	Ira Weiny <ira.weiny@...el.com>, linux-rdma@...r.kernel.org,
	linux-kernel@...r.kernel.org, linux-nfs@...r.kernel.org,
	netdev@...r.kernel.org, "J. Bruce Fields" <bfields@...ldses.org>,
	Trond Myklebust <trond.myklebust@...marydata.com>,
	"David S. Miller" <davem@...emloft.net>,
	Or Gerlitz <ogerlitz@...lanox.com>,
	Moni Shoua <monis@...lanox.com>,
	PJ Waskiewicz <pj.waskiewicz@...idfire.com>,
	Tatyana Nikolova <Tatyana.E.Nikolova@...el.com>,
	Yan Burman <yanb@...lanox.com>,
	Jack Morgenstein <jackm@....mellanox.co.il>,
	Bart Van Assche <bvanassche@....org>,
	Yann Droneaud <ydroneaud@...eya.com>,
	Colin Ian King <colin.king@...onical.com>,
	Majd Dibbiny <majd@...lanox.com>,
	Jiri Kosina <jkosina@...e.cz>,
	Matan Barak <matanb@...lanox.com>,
	Alex Estrin <alex.estrin@...el.com>,
	Eric Dumazet <edumazet@...gle.com>,
	Erez Shitrit <erezsh@...lanox.com>,
	Sagi Grimberg <sagig@...lanox.com>,
	Haggai Eran <haggaie@...lanox.com>,
	Shachar Raindel <raindel@...lanox.com>,
	Mike Marciniszyn <mike.marciniszyn@...el.com>,
	Steve Wise <swise@...ngridcomputing.com>,
	Tom Tucker <tom@....us>, Chuck Lever <chuck.lever@...cle.com>
Subject: Re: [PATCH 01/11] IB/Verbs: Use helpers to check transport and link
 layer

On 03/30/2015 06:22 PM, Doug Ledford wrote:
> On Mon, 2015-03-30 at 18:14 +0200, Michael Wang wrote:
>> [snip]
> There is no "gradually eliminate them" to the suggestion I made.
> Remember, my suggestion was to remove the transport and link_layer items
> from the port settings and replace it with just one transport item that
> is a bitmask of the possible transport types.  This can not be done
> gradually, it must be a complete change all at once as the two methods
> of setting things are incompatible.  As there is only one out of tree
> driver that I know of, lustre, we can give them the information they
> need to make their driver work both before and after the change.

Actually there is something confused me on transport and link
layer here, basically we have defined:

transport type
        RDMA_TRANSPORT_IB,
        RDMA_TRANSPORT_IWARP,
        RDMA_TRANSPORT_USNIC,
        RDMA_TRANSPORT_USNIC_UDP
link layer
        IB_LINK_LAYER_INFINIBAND,
        IB_LINK_LAYER_ETHERNET,

So we could have a table:

                                            LL_INFINIBAND    LL_ETHERNET    UNCARE
TRANSPORT_IB                    1                            2                        3
TRANSPORT_IWARP,                                                                    4
UNCARE                               5                            6                      

In current implementation I've found all these combination
in core or driver, and I could see:

rdma_transport_is_ib()		1
rdma_transport_is_iwarp()	4	
rdma_transport_is_roce()	2

Just confusing how to take care the combination 3,5,6?

Regards,
Michael Wang

>
>> Sure if finally we do capture all the cases, we can just get rid of
>> this one, but I guess it won't be that easy to directly jump into
>> next stage :-P
>>
>> As I could imaging, after this reform, next stage could be introducing
>> the new mechanism without changing device driver, and the last
>> stage is to asking vendor adapt their code into the new mechanism.
>>
>>> In other words, if our end goal is to have
>>>
>>> rdma_transport_is_ib()
>>> rdma_transport_is_iwarp()
>>> rdma_transport_is_roce()
>>> rdma_transport_is_opa()
>>>
>>> Then we should skip doing rdma_port_ll_is_*() as the answers to these
>>> items would be implied by rdma_transport_is_roce() and such.
>> Great if we achieved that ;-) but currently I just wondering maybe
>> these helpers can only cover part of the cases where we check
>> transport and link layer, there are still some cases we'll need the
>> very rough helper to save some code and make things clean~
>>
>> Regards,
>> Michael Wang
>>
>>
>>>> Cc: Jason Gunthorpe <jgunthorpe@...idianresearch.com>
>>>> Cc: Doug Ledford <dledford@...hat.com>
>>>> Cc: Ira Weiny <ira.weiny@...el.com>
>>>> Cc: Sean Hefty <sean.hefty@...el.com>
>>>> Signed-off-by: Michael Wang <yun.wang@...fitbricks.com>
>>>> ---
>>>>  drivers/infiniband/core/agent.c           |  2 +-
>>>>  drivers/infiniband/core/cm.c              |  2 +-
>>>>  drivers/infiniband/core/cma.c             | 27 ++++++++++++---------------
>>>>  drivers/infiniband/core/mad.c             |  6 +++---
>>>>  drivers/infiniband/core/multicast.c       | 11 ++++-------
>>>>  drivers/infiniband/core/sa_query.c        | 14 +++++++-------
>>>>  drivers/infiniband/core/ucm.c             |  3 +--
>>>>  drivers/infiniband/core/user_mad.c        |  2 +-
>>>>  drivers/infiniband/core/verbs.c           |  5 ++---
>>>>  drivers/infiniband/hw/mlx4/ah.c           |  2 +-
>>>>  drivers/infiniband/hw/mlx4/cq.c           |  4 +---
>>>>  drivers/infiniband/hw/mlx4/mad.c          | 14 ++++----------
>>>>  drivers/infiniband/hw/mlx4/main.c         |  8 +++-----
>>>>  drivers/infiniband/hw/mlx4/mlx4_ib.h      |  2 +-
>>>>  drivers/infiniband/hw/mlx4/qp.c           | 21 +++++++--------------
>>>>  drivers/infiniband/hw/mlx4/sysfs.c        |  6 ++----
>>>>  drivers/infiniband/ulp/ipoib/ipoib_main.c |  6 +++---
>>>>  include/rdma/ib_verbs.h                   | 24 ++++++++++++++++++++++++
>>>>  net/sunrpc/xprtrdma/svc_rdma_recvfrom.c   |  3 +--
>>>>  19 files changed, 79 insertions(+), 83 deletions(-)
>>>>
>>>> diff --git a/drivers/infiniband/core/agent.c b/drivers/infiniband/core/agent.c
>>>> index f6d2961..27f1bec 100644
>>>> --- a/drivers/infiniband/core/agent.c
>>>> +++ b/drivers/infiniband/core/agent.c
>>>> @@ -156,7 +156,7 @@ int ib_agent_port_open(struct ib_device *device, int port_num)
>>>>          goto error1;
>>>>      }
>>>>  
>>>> -    if (rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_INFINIBAND) {
>>>> +    if (rdma_port_ll_is_ib(device, port_num)) {
>>>>          /* Obtain send only MAD agent for SMI QP */
>>>>          port_priv->agent[0] = ib_register_mad_agent(device, port_num,
>>>>                                  IB_QPT_SMI, NULL, 0,
>>>> diff --git a/drivers/infiniband/core/cm.c b/drivers/infiniband/core/cm.c
>>>> index e28a494..2c72e9e 100644
>>>> --- a/drivers/infiniband/core/cm.c
>>>> +++ b/drivers/infiniband/core/cm.c
>>>> @@ -3762,7 +3762,7 @@ static void cm_add_one(struct ib_device *ib_device)
>>>>      int ret;
>>>>      u8 i;
>>>>  
>>>> -    if (rdma_node_get_transport(ib_device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(ib_device))
>>>>          return;
>>>>  
>>>>      cm_dev = kzalloc(sizeof(*cm_dev) + sizeof(*port) *
>>>> diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c
>>>> index d570030..668e955 100644
>>>> --- a/drivers/infiniband/core/cma.c
>>>> +++ b/drivers/infiniband/core/cma.c
>>>> @@ -375,8 +375,8 @@ static int cma_acquire_dev(struct rdma_id_private *id_priv,
>>>>                       listen_id_priv->id.port_num) == dev_ll) {
>>>>          cma_dev = listen_id_priv->cma_dev;
>>>>          port = listen_id_priv->id.port_num;
>>>> -        if (rdma_node_get_transport(cma_dev->device->node_type) == RDMA_TRANSPORT_IB &&
>>>> -            rdma_port_get_link_layer(cma_dev->device, port) == IB_LINK_LAYER_ETHERNET)
>>>> +        if (rdma_transport_is_ib(cma_dev->device) &&
>>>> +            rdma_port_ll_is_eth(cma_dev->device, port))
>>>>              ret = ib_find_cached_gid(cma_dev->device, &iboe_gid,
>>>>                           &found_port, NULL);
>>>>          else
>>>> @@ -395,8 +395,8 @@ static int cma_acquire_dev(struct rdma_id_private *id_priv,
>>>>                  listen_id_priv->id.port_num == port)
>>>>                  continue;
>>>>              if (rdma_port_get_link_layer(cma_dev->device, port) == dev_ll) {
>>>> -                if (rdma_node_get_transport(cma_dev->device->node_type) == RDMA_TRANSPORT_IB &&
>>>> -                    rdma_port_get_link_layer(cma_dev->device, port) == IB_LINK_LAYER_ETHERNET)
>>>> +                if (rdma_transport_is_ib(cma_dev->device) &&
>>>> +                    rdma_port_ll_is_eth(cma_dev->device, port))
>>>>                      ret = ib_find_cached_gid(cma_dev->device, &iboe_gid, &found_port, NULL);
>>>>                  else
>>>>                      ret = ib_find_cached_gid(cma_dev->device, &gid, &found_port, NULL);
>>>> @@ -435,7 +435,7 @@ static int cma_resolve_ib_dev(struct rdma_id_private *id_priv)
>>>>      pkey = ntohs(addr->sib_pkey);
>>>>  
>>>>      list_for_each_entry(cur_dev, &dev_list, list) {
>>>> -        if (rdma_node_get_transport(cur_dev->device->node_type) != RDMA_TRANSPORT_IB)
>>>> +        if (!rdma_transport_is_ib(cur_dev->device))
>>>>              continue;
>>>>  
>>>>          for (p = 1; p <= cur_dev->device->phys_port_cnt; ++p) {
>>>> @@ -633,10 +633,8 @@ static int cma_modify_qp_rtr(struct rdma_id_private *id_priv,
>>>>      if (ret)
>>>>          goto out;
>>>>  
>>>> -    if (rdma_node_get_transport(id_priv->cma_dev->device->node_type)
>>>> -        == RDMA_TRANSPORT_IB &&
>>>> -        rdma_port_get_link_layer(id_priv->id.device, id_priv->id.port_num)
>>>> -        == IB_LINK_LAYER_ETHERNET) {
>>>> +    if (rdma_transport_is_ib(id_priv->cma_dev->device) &&
>>>> +        rdma_port_ll_is_eth(id_priv->id.device, id_priv->id.port_num)) {
>>>>          ret = rdma_addr_find_smac_by_sgid(&sgid, qp_attr.smac, NULL);
>>>>  
>>>>          if (ret)
>>>> @@ -700,8 +698,7 @@ static int cma_ib_init_qp_attr(struct rdma_id_private *id_priv,
>>>>      int ret;
>>>>      u16 pkey;
>>>>  
>>>> -    if (rdma_port_get_link_layer(id_priv->id.device, id_priv->id.port_num) ==
>>>> -        IB_LINK_LAYER_INFINIBAND)
>>>> +    if (rdma_port_ll_is_ib(id_priv->id.device, id_priv->id.port_num))
>>>>          pkey = ib_addr_get_pkey(dev_addr);
>>>>      else
>>>>          pkey = 0xffff;
>>>> @@ -1626,7 +1623,7 @@ static void cma_listen_on_dev(struct rdma_id_private *id_priv,
>>>>      int ret;
>>>>  
>>>>      if (cma_family(id_priv) == AF_IB &&
>>>> -        rdma_node_get_transport(cma_dev->device->node_type) != RDMA_TRANSPORT_IB)
>>>> +        !rdma_transport_is_ib(cma_dev->device))
>>>>          return;
>>>>  
>>>>      id = rdma_create_id(cma_listen_handler, id_priv, id_priv->id.ps,
>>>> @@ -2028,7 +2025,7 @@ static int cma_bind_loopback(struct rdma_id_private *id_priv)
>>>>      mutex_lock(&lock);
>>>>      list_for_each_entry(cur_dev, &dev_list, list) {
>>>>          if (cma_family(id_priv) == AF_IB &&
>>>> -            rdma_node_get_transport(cur_dev->device->node_type) != RDMA_TRANSPORT_IB)
>>>> +            !rdma_transport_is_ib(cur_dev->device))
>>>>              continue;
>>>>  
>>>>          if (!cma_dev)
>>>> @@ -2060,7 +2057,7 @@ port_found:
>>>>          goto out;
>>>>  
>>>>      id_priv->id.route.addr.dev_addr.dev_type =
>>>> -        (rdma_port_get_link_layer(cma_dev->device, p) == IB_LINK_LAYER_INFINIBAND) ?
>>>> +        (rdma_port_ll_is_ib(cma_dev->device, p)) ?
>>>>          ARPHRD_INFINIBAND : ARPHRD_ETHER;
>>>>  
>>>>      rdma_addr_set_sgid(&id_priv->id.route.addr.dev_addr, &gid);
>>>> @@ -3405,7 +3402,7 @@ void rdma_leave_multicast(struct rdma_cm_id *id, struct sockaddr *addr)
>>>>                  ib_detach_mcast(id->qp,
>>>>                          &mc->multicast.ib->rec.mgid,
>>>>                          be16_to_cpu(mc->multicast.ib->rec.mlid));
>>>> -            if (rdma_node_get_transport(id_priv->cma_dev->device->node_type) == RDMA_TRANSPORT_IB) {
>>>> +            if (rdma_transport_is_ib(id_priv->cma_dev->device)) {
>>>>                  switch (rdma_port_get_link_layer(id->device, id->port_num)) {
>>>>                  case IB_LINK_LAYER_INFINIBAND:
>>>>                      ib_sa_free_multicast(mc->multicast.ib);
>>>> diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c
>>>> index 74c30f4..23cf9e8 100644
>>>> --- a/drivers/infiniband/core/mad.c
>>>> +++ b/drivers/infiniband/core/mad.c
>>>> @@ -2938,7 +2938,7 @@ static int ib_mad_port_open(struct ib_device *device,
>>>>      init_mad_qp(port_priv, &port_priv->qp_info[1]);
>>>>  
>>>>      cq_size = mad_sendq_size + mad_recvq_size;
>>>> -    has_smi = rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_INFINIBAND;
>>>> +    has_smi = rdma_port_ll_is_ib(device, port_num);
>>>>      if (has_smi)
>>>>          cq_size *= 2;
>>>>  
>>>> @@ -3057,7 +3057,7 @@ static void ib_mad_init_device(struct ib_device *device)
>>>>  {
>>>>      int start, end, i;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      if (device->node_type == RDMA_NODE_IB_SWITCH) {
>>>> @@ -3102,7 +3102,7 @@ static void ib_mad_remove_device(struct ib_device *device)
>>>>  {
>>>>      int i, num_ports, cur_port;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      if (device->node_type == RDMA_NODE_IB_SWITCH) {
>>>> diff --git a/drivers/infiniband/core/multicast.c b/drivers/infiniband/core/multicast.c
>>>> index fa17b55..17573ff 100644
>>>> --- a/drivers/infiniband/core/multicast.c
>>>> +++ b/drivers/infiniband/core/multicast.c
>>>> @@ -780,8 +780,7 @@ static void mcast_event_handler(struct ib_event_handler *handler,
>>>>      int index;
>>>>  
>>>>      dev = container_of(handler, struct mcast_device, event_handler);
>>>> -    if (rdma_port_get_link_layer(dev->device, event->element.port_num) !=
>>>> -        IB_LINK_LAYER_INFINIBAND)
>>>> +    if (!rdma_port_ll_is_ib(dev->device, event->element.port_num))
>>>>          return;
>>>>  
>>>>      index = event->element.port_num - dev->start_port;
>>>> @@ -808,7 +807,7 @@ static void mcast_add_one(struct ib_device *device)
>>>>      int i;
>>>>      int count = 0;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      dev = kmalloc(sizeof *dev + device->phys_port_cnt * sizeof *port,
>>>> @@ -824,8 +823,7 @@ static void mcast_add_one(struct ib_device *device)
>>>>      }
>>>>  
>>>>      for (i = 0; i <= dev->end_port - dev->start_port; i++) {
>>>> -        if (rdma_port_get_link_layer(device, dev->start_port + i) !=
>>>> -            IB_LINK_LAYER_INFINIBAND)
>>>> +        if (!rdma_port_ll_is_ib(device, dev->start_port + i))
>>>>              continue;
>>>>          port = &dev->port[i];
>>>>          port->dev = dev;
>>>> @@ -863,8 +861,7 @@ static void mcast_remove_one(struct ib_device *device)
>>>>      flush_workqueue(mcast_wq);
>>>>  
>>>>      for (i = 0; i <= dev->end_port - dev->start_port; i++) {
>>>> -        if (rdma_port_get_link_layer(device, dev->start_port + i) ==
>>>> -            IB_LINK_LAYER_INFINIBAND) {
>>>> +        if (rdma_port_ll_is_ib(device, dev->start_port + i)) {
>>>>              port = &dev->port[i];
>>>>              deref_port(port);
>>>>              wait_for_completion(&port->comp);
>>>> diff --git a/drivers/infiniband/core/sa_query.c b/drivers/infiniband/core/sa_query.c
>>>> index c38f030..d95d25f 100644
>>>> --- a/drivers/infiniband/core/sa_query.c
>>>> +++ b/drivers/infiniband/core/sa_query.c
>>>> @@ -450,7 +450,7 @@ static void ib_sa_event(struct ib_event_handler *handler, struct ib_event *event
>>>>          struct ib_sa_port *port =
>>>>              &sa_dev->port[event->element.port_num - sa_dev->start_port];
>>>>  
>>>> -        if (rdma_port_get_link_layer(handler->device, port->port_num) != IB_LINK_LAYER_INFINIBAND)
>>>> +        if (!rdma_port_ll_is_ib(handler->device, port->port_num))
>>>>              return;
>>>>  
>>>>          spin_lock_irqsave(&port->ah_lock, flags);
>>>> @@ -540,7 +540,7 @@ int ib_init_ah_from_path(struct ib_device *device, u8 port_num,
>>>>      ah_attr->port_num = port_num;
>>>>      ah_attr->static_rate = rec->rate;
>>>>  
>>>> -    force_grh = rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_ETHERNET;
>>>> +    force_grh = rdma_port_ll_is_eth(device, port_num);
>>>>  
>>>>      if (rec->hop_limit > 1 || force_grh) {
>>>>          ah_attr->ah_flags = IB_AH_GRH;
>>>> @@ -1154,7 +1154,7 @@ static void ib_sa_add_one(struct ib_device *device)
>>>>      struct ib_sa_device *sa_dev;
>>>>      int s, e, i;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      if (device->node_type == RDMA_NODE_IB_SWITCH)
>>>> @@ -1175,7 +1175,7 @@ static void ib_sa_add_one(struct ib_device *device)
>>>>  
>>>>      for (i = 0; i <= e - s; ++i) {
>>>>          spin_lock_init(&sa_dev->port[i].ah_lock);
>>>> -        if (rdma_port_get_link_layer(device, i + 1) != IB_LINK_LAYER_INFINIBAND)
>>>> +        if (!rdma_port_ll_is_ib(device, i + 1))
>>>>              continue;
>>>>  
>>>>          sa_dev->port[i].sm_ah    = NULL;
>>>> @@ -1205,14 +1205,14 @@ static void ib_sa_add_one(struct ib_device *device)
>>>>          goto err;
>>>>  
>>>>      for (i = 0; i <= e - s; ++i)
>>>> -        if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND)
>>>> +        if (rdma_port_ll_is_ib(device, i + 1))
>>>>              update_sm_ah(&sa_dev->port[i].update_task);
>>>>  
>>>>      return;
>>>>  
>>>>  err:
>>>>      while (--i >= 0)
>>>> -        if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND)
>>>> +        if (rdma_port_ll_is_ib(device, i + 1))
>>>>              ib_unregister_mad_agent(sa_dev->port[i].agent);
>>>>  
>>>>      kfree(sa_dev);
>>>> @@ -1233,7 +1233,7 @@ static void ib_sa_remove_one(struct ib_device *device)
>>>>      flush_workqueue(ib_wq);
>>>>  
>>>>      for (i = 0; i <= sa_dev->end_port - sa_dev->start_port; ++i) {
>>>> -        if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND) {
>>>> +        if (rdma_port_ll_is_ib(device, i + 1)) {
>>>>              ib_unregister_mad_agent(sa_dev->port[i].agent);
>>>>              if (sa_dev->port[i].sm_ah)
>>>>                  kref_put(&sa_dev->port[i].sm_ah->ref, free_sm_ah);
>>>> diff --git a/drivers/infiniband/core/ucm.c b/drivers/infiniband/core/ucm.c
>>>> index f2f6393..ddbe0b4 100644
>>>> --- a/drivers/infiniband/core/ucm.c
>>>> +++ b/drivers/infiniband/core/ucm.c
>>>> @@ -1253,8 +1253,7 @@ static void ib_ucm_add_one(struct ib_device *device)
>>>>      dev_t base;
>>>>      struct ib_ucm_device *ucm_dev;
>>>>  
>>>> -    if (!device->alloc_ucontext ||
>>>> -        rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!device->alloc_ucontext || !rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      ucm_dev = kzalloc(sizeof *ucm_dev, GFP_KERNEL);
>>>> diff --git a/drivers/infiniband/core/user_mad.c b/drivers/infiniband/core/user_mad.c
>>>> index 928cdd2..28a8b30 100644
>>>> --- a/drivers/infiniband/core/user_mad.c
>>>> +++ b/drivers/infiniband/core/user_mad.c
>>>> @@ -1274,7 +1274,7 @@ static void ib_umad_add_one(struct ib_device *device)
>>>>      struct ib_umad_device *umad_dev;
>>>>      int s, e, i;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      if (device->node_type == RDMA_NODE_IB_SWITCH)
>>>> diff --git a/drivers/infiniband/core/verbs.c b/drivers/infiniband/core/verbs.c
>>>> index f93eb8d..d8d015a 100644
>>>> --- a/drivers/infiniband/core/verbs.c
>>>> +++ b/drivers/infiniband/core/verbs.c
>>>> @@ -198,8 +198,7 @@ int ib_init_ah_from_wc(struct ib_device *device, u8 port_num, struct ib_wc *wc,
>>>>      u32 flow_class;
>>>>      u16 gid_index;
>>>>      int ret;
>>>> -    int is_eth = (rdma_port_get_link_layer(device, port_num) ==
>>>> -            IB_LINK_LAYER_ETHERNET);
>>>> +    int is_eth = (rdma_port_ll_is_eth(device, port_num));
>>>>  
>>>>      memset(ah_attr, 0, sizeof *ah_attr);
>>>>      if (is_eth) {
>>>> @@ -871,7 +870,7 @@ int ib_resolve_eth_l2_attrs(struct ib_qp *qp,
>>>>      union ib_gid  sgid;
>>>>  
>>>>      if ((*qp_attr_mask & IB_QP_AV)  &&
>>>> -        (rdma_port_get_link_layer(qp->device, qp_attr->ah_attr.port_num) == IB_LINK_LAYER_ETHERNET)) {
>>>> +        (rdma_port_ll_is_eth(qp->device, qp_attr->ah_attr.port_num))) {
>>>>          ret = ib_query_gid(qp->device, qp_attr->ah_attr.port_num,
>>>>                     qp_attr->ah_attr.grh.sgid_index, &sgid);
>>>>          if (ret)
>>>> diff --git a/drivers/infiniband/hw/mlx4/ah.c b/drivers/infiniband/hw/mlx4/ah.c
>>>> index 2d8c339..829eb60 100644
>>>> --- a/drivers/infiniband/hw/mlx4/ah.c
>>>> +++ b/drivers/infiniband/hw/mlx4/ah.c
>>>> @@ -118,7 +118,7 @@ struct ib_ah *mlx4_ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
>>>>      if (!ah)
>>>>          return ERR_PTR(-ENOMEM);
>>>>  
>>>> -    if (rdma_port_get_link_layer(pd->device, ah_attr->port_num) == IB_LINK_LAYER_ETHERNET) {
>>>> +    if (rdma_port_ll_is_eth(pd->device, ah_attr->port_num)) {
>>>>          if (!(ah_attr->ah_flags & IB_AH_GRH)) {
>>>>              ret = ERR_PTR(-EINVAL);
>>>>          } else {
>>>> diff --git a/drivers/infiniband/hw/mlx4/cq.c b/drivers/infiniband/hw/mlx4/cq.c
>>>> index cb63ecd..0417f03 100644
>>>> --- a/drivers/infiniband/hw/mlx4/cq.c
>>>> +++ b/drivers/infiniband/hw/mlx4/cq.c
>>>> @@ -789,9 +789,7 @@ repoll:
>>>>              break;
>>>>          }
>>>>  
>>>> -        is_eth = (rdma_port_get_link_layer(wc->qp->device,
>>>> -                          (*cur_qp)->port) ==
>>>> -              IB_LINK_LAYER_ETHERNET);
>>>> +        is_eth = (rdma_port_ll_is_eth(wc->qp->device, (*cur_qp)->port));
>>>>          if (mlx4_is_mfunc(to_mdev(cq->ibcq.device)->dev)) {
>>>>              if ((*cur_qp)->mlx4_ib_qp_type &
>>>>                  (MLX4_IB_QPT_PROXY_SMI_OWNER |
>>>> diff --git a/drivers/infiniband/hw/mlx4/mad.c b/drivers/infiniband/hw/mlx4/mad.c
>>>> index 82a7dd8..4736fc7 100644
>>>> --- a/drivers/infiniband/hw/mlx4/mad.c
>>>> +++ b/drivers/infiniband/hw/mlx4/mad.c
>>>> @@ -606,12 +606,7 @@ static int mlx4_ib_demux_mad(struct ib_device *ibdev, u8 port,
>>>>      int err;
>>>>      int slave;
>>>>      u8 *slave_id;
>>>> -    int is_eth = 0;
>>>> -
>>>> -    if (rdma_port_get_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND)
>>>> -        is_eth = 0;
>>>> -    else
>>>> -        is_eth = 1;
>>>> +    int is_eth = rdma_port_ll_is_eth(ibdev, port);
>>>>  
>>>>      if (is_eth) {
>>>>          if (!(wc->wc_flags & IB_WC_GRH)) {
>>>> @@ -1252,7 +1247,7 @@ out:
>>>>  
>>>>  static int get_slave_base_gid_ix(struct mlx4_ib_dev *dev, int slave, int port)
>>>>  {
>>>> -    if (rdma_port_get_link_layer(&dev->ib_dev, port) == IB_LINK_LAYER_INFINIBAND)
>>>> +    if (rdma_port_ll_is_ib(&dev->ib_dev, port))
>>>>          return slave;
>>>>      return mlx4_get_base_gid_ix(dev->dev, slave, port);
>>>>  }
>>>> @@ -1260,7 +1255,7 @@ static int get_slave_base_gid_ix(struct mlx4_ib_dev *dev, int slave, int port)
>>>>  static void fill_in_real_sgid_index(struct mlx4_ib_dev *dev, int slave, int port,
>>>>                      struct ib_ah_attr *ah_attr)
>>>>  {
>>>> -    if (rdma_port_get_link_layer(&dev->ib_dev, port) == IB_LINK_LAYER_INFINIBAND)
>>>> +    if (rdma_port_ll_is_ib(&dev->ib_dev, port))
>>>>          ah_attr->grh.sgid_index = slave;
>>>>      else
>>>>          ah_attr->grh.sgid_index += get_slave_base_gid_ix(dev, slave, port);
>>>> @@ -1758,8 +1753,7 @@ static int create_pv_resources(struct ib_device *ibdev, int slave, int port,
>>>>  
>>>>      ctx->state = DEMUX_PV_STATE_STARTING;
>>>>      /* have QP0 only if link layer is IB */
>>>> -    if (rdma_port_get_link_layer(ibdev, ctx->port) ==
>>>> -        IB_LINK_LAYER_INFINIBAND)
>>>> +    if (rdma_port_ll_is_ib(ibdev, ctx->port))
>>>>          ctx->has_smi = 1;
>>>>  
>>>>      if (ctx->has_smi) {
>>>> diff --git a/drivers/infiniband/hw/mlx4/main.c b/drivers/infiniband/hw/mlx4/main.c
>>>> index 0b280b1..f445f4c 100644
>>>> --- a/drivers/infiniband/hw/mlx4/main.c
>>>> +++ b/drivers/infiniband/hw/mlx4/main.c
>>>> @@ -482,7 +482,7 @@ static int iboe_query_gid(struct ib_device *ibdev, u8 port, int index,
>>>>  static int mlx4_ib_query_gid(struct ib_device *ibdev, u8 port, int index,
>>>>                   union ib_gid *gid)
>>>>  {
>>>> -    if (rdma_port_get_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND)
>>>> +    if (rdma_port_ll_is_ib(ibdev, port))
>>>>          return __mlx4_ib_query_gid(ibdev, port, index, gid, 0);
>>>>      else
>>>>          return iboe_query_gid(ibdev, port, index, gid);
>>>> @@ -1801,8 +1801,7 @@ static int mlx4_ib_init_gid_table(struct mlx4_ib_dev *ibdev)
>>>>      int err = 0;
>>>>  
>>>>      for (i = 1; i <= ibdev->num_ports; ++i) {
>>>> -        if (rdma_port_get_link_layer(&ibdev->ib_dev, i) ==
>>>> -            IB_LINK_LAYER_ETHERNET) {
>>>> +        if (rdma_port_ll_is_eth(&ibdev->ib_dev, i)) {
>>>>              err = reset_gid_table(ibdev, i);
>>>>              if (err)
>>>>                  goto out;
>>>> @@ -2554,8 +2553,7 @@ static void mlx4_ib_event(struct mlx4_dev *dev, void *ibdev_ptr,
>>>>          if (p > ibdev->num_ports)
>>>>              return;
>>>>          if (mlx4_is_master(dev) &&
>>>> -            rdma_port_get_link_layer(&ibdev->ib_dev, p) ==
>>>> -            IB_LINK_LAYER_INFINIBAND) {
>>>> +            rdma_port_ll_is_ib(&ibdev->ib_dev, p)) {
>>>>              mlx4_ib_invalidate_all_guid_record(ibdev, p);
>>>>          }
>>>>          ibev.event = IB_EVENT_PORT_ACTIVE;
>>>> diff --git a/drivers/infiniband/hw/mlx4/mlx4_ib.h b/drivers/infiniband/hw/mlx4/mlx4_ib.h
>>>> index 6eb743f..1befeb8 100644
>>>> --- a/drivers/infiniband/hw/mlx4/mlx4_ib.h
>>>> +++ b/drivers/infiniband/hw/mlx4/mlx4_ib.h
>>>> @@ -712,7 +712,7 @@ static inline bool mlx4_ib_ah_grh_present(struct mlx4_ib_ah *ah)
>>>>  {
>>>>      u8 port = be32_to_cpu(ah->av.ib.port_pd) >> 24 & 3;
>>>>  
>>>> -    if (rdma_port_get_link_layer(ah->ibah.device, port) == IB_LINK_LAYER_ETHERNET)
>>>> +    if (rdma_port_ll_is_eth(ah->ibah.device, port))
>>>>          return true;
>>>>  
>>>>      return !!(ah->av.ib.g_slid & 0x80);
>>>> diff --git a/drivers/infiniband/hw/mlx4/qp.c b/drivers/infiniband/hw/mlx4/qp.c
>>>> index c880329..bd2f557 100644
>>>> --- a/drivers/infiniband/hw/mlx4/qp.c
>>>> +++ b/drivers/infiniband/hw/mlx4/qp.c
>>>> @@ -1248,8 +1248,7 @@ static int _mlx4_set_path(struct mlx4_ib_dev *dev, const struct ib_ah_attr *ah,
>>>>                u64 smac, u16 vlan_tag, struct mlx4_qp_path *path,
>>>>                struct mlx4_roce_smac_vlan_info *smac_info, u8 port)
>>>>  {
>>>> -    int is_eth = rdma_port_get_link_layer(&dev->ib_dev, port) ==
>>>> -        IB_LINK_LAYER_ETHERNET;
>>>> +    int is_eth = rdma_port_ll_is_eth(&dev->ib_dev, port);
>>>>      int vidx;
>>>>      int smac_index;
>>>>      int err;
>>>> @@ -1433,8 +1432,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
>>>>  
>>>>      /* APM is not supported under RoCE */
>>>>      if (attr_mask & IB_QP_ALT_PATH &&
>>>> -        rdma_port_get_link_layer(&dev->ib_dev, qp->port) ==
>>>> -        IB_LINK_LAYER_ETHERNET)
>>>> +        rdma_port_ll_is_eth(&dev->ib_dev, qp->port))
>>>>          return -ENOTSUPP;
>>>>  
>>>>      context = kzalloc(sizeof *context, GFP_KERNEL);
>>>> @@ -1664,8 +1662,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
>>>>                  context->pri_path.fl = 0x80;
>>>>              context->pri_path.sched_queue |= MLX4_IB_DEFAULT_SCHED_QUEUE;
>>>>          }
>>>> -        if (rdma_port_get_link_layer(&dev->ib_dev, qp->port) ==
>>>> -            IB_LINK_LAYER_ETHERNET) {
>>>> +        if (rdma_port_ll_is_eth(&dev->ib_dev, qp->port)) {
>>>>              if (qp->mlx4_ib_qp_type == MLX4_IB_QPT_TUN_GSI ||
>>>>                  qp->mlx4_ib_qp_type == MLX4_IB_QPT_GSI)
>>>>                  context->pri_path.feup = 1 << 7; /* don't fsm */
>>>> @@ -1695,9 +1692,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
>>>>      }
>>>>  
>>>>      if (ibqp->qp_type == IB_QPT_UD && (new_state == IB_QPS_RTR)) {
>>>> -        int is_eth = rdma_port_get_link_layer(
>>>> -                &dev->ib_dev, qp->port) ==
>>>> -                IB_LINK_LAYER_ETHERNET;
>>>> +        int is_eth = rdma_port_ll_is_eth(&dev->ib_dev, qp->port);
>>>>          if (is_eth) {
>>>>              context->pri_path.ackto = MLX4_IB_LINK_TYPE_ETH;
>>>>              optpar |= MLX4_QP_OPTPAR_PRIMARY_ADDR_PATH;
>>>> @@ -1927,8 +1922,7 @@ int mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
>>>>      }
>>>>  
>>>>      if ((attr_mask & IB_QP_PORT) && (ibqp->qp_type == IB_QPT_RAW_PACKET) &&
>>>> -        (rdma_port_get_link_layer(&dev->ib_dev, attr->port_num) !=
>>>> -         IB_LINK_LAYER_ETHERNET))
>>>> +        !rdma_port_ll_is_eth(&dev->ib_dev, attr->port_num))
>>>>          goto out;
>>>>  
>>>>      if (attr_mask & IB_QP_PKEY_INDEX) {
>>>> @@ -2132,7 +2126,7 @@ static int build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr,
>>>>      for (i = 0; i < wr->num_sge; ++i)
>>>>          send_size += wr->sg_list[i].length;
>>>>  
>>>> -    is_eth = rdma_port_get_link_layer(sqp->qp.ibqp.device, sqp->qp.port) == IB_LINK_LAYER_ETHERNET;
>>>> +    is_eth = rdma_port_ll_is_eth(sqp->qp.ibqp.device, sqp->qp.port);
>>>>      is_grh = mlx4_ib_ah_grh_present(ah);
>>>>      if (is_eth) {
>>>>          if (mlx4_is_mfunc(to_mdev(ib_dev)->dev)) {
>>>> @@ -3029,8 +3023,7 @@ static void to_ib_ah_attr(struct mlx4_ib_dev *ibdev, struct ib_ah_attr *ib_ah_at
>>>>      if (ib_ah_attr->port_num == 0 || ib_ah_attr->port_num > dev->caps.num_ports)
>>>>          return;
>>>>  
>>>> -    is_eth = rdma_port_get_link_layer(&ibdev->ib_dev, ib_ah_attr->port_num) ==
>>>> -        IB_LINK_LAYER_ETHERNET;
>>>> +    is_eth = rdma_port_ll_is_eth(&ibdev->ib_dev, ib_ah_attr->port_num);
>>>>      if (is_eth)
>>>>          ib_ah_attr->sl = ((path->sched_queue >> 3) & 0x7) |
>>>>          ((path->sched_queue & 4) << 1);
>>>> diff --git a/drivers/infiniband/hw/mlx4/sysfs.c b/drivers/infiniband/hw/mlx4/sysfs.c
>>>> index cb4c66e..d339b55 100644
>>>> --- a/drivers/infiniband/hw/mlx4/sysfs.c
>>>> +++ b/drivers/infiniband/hw/mlx4/sysfs.c
>>>> @@ -610,8 +610,7 @@ static ssize_t sysfs_store_enable_smi_admin(struct device *dev,
>>>>  
>>>>  static int add_vf_smi_entries(struct mlx4_port *p)
>>>>  {
>>>> -    int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) ==
>>>> -            IB_LINK_LAYER_ETHERNET;
>>>> +    int is_eth = rdma_port_ll_is_eth(&p->dev->ib_dev, p->port_num);
>>>>      int ret;
>>>>  
>>>>      /* do not display entries if eth transport, or if master */
>>>> @@ -645,8 +644,7 @@ static int add_vf_smi_entries(struct mlx4_port *p)
>>>>  
>>>>  static void remove_vf_smi_entries(struct mlx4_port *p)
>>>>  {
>>>> -    int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) ==
>>>> -            IB_LINK_LAYER_ETHERNET;
>>>> +    int is_eth = rdma_port_ll_is_eth(&p->dev->ib_dev, p->port_num);
>>>>  
>>>>      if (is_eth || p->slave == mlx4_master_func_num(p->dev->dev))
>>>>          return;
>>>> diff --git a/drivers/infiniband/ulp/ipoib/ipoib_main.c b/drivers/infiniband/ulp/ipoib/ipoib_main.c
>>>> index 58b5aa3..3341754 100644
>>>> --- a/drivers/infiniband/ulp/ipoib/ipoib_main.c
>>>> +++ b/drivers/infiniband/ulp/ipoib/ipoib_main.c
>>>> @@ -1655,7 +1655,7 @@ static void ipoib_add_one(struct ib_device *device)
>>>>      struct ipoib_dev_priv *priv;
>>>>      int s, e, p;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      dev_list = kmalloc(sizeof *dev_list, GFP_KERNEL);
>>>> @@ -1673,7 +1673,7 @@ static void ipoib_add_one(struct ib_device *device)
>>>>      }
>>>>  
>>>>      for (p = s; p <= e; ++p) {
>>>> -        if (rdma_port_get_link_layer(device, p) != IB_LINK_LAYER_INFINIBAND)
>>>> +        if (!rdma_port_ll_is_ib(device, p))
>>>>              continue;
>>>>          dev = ipoib_add_port("ib%d", device, p);
>>>>          if (!IS_ERR(dev)) {
>>>> @@ -1690,7 +1690,7 @@ static void ipoib_remove_one(struct ib_device *device)
>>>>      struct ipoib_dev_priv *priv, *tmp;
>>>>      struct list_head *dev_list;
>>>>  
>>>> -    if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
>>>> +    if (!rdma_transport_is_ib(device))
>>>>          return;
>>>>  
>>>>      dev_list = ib_get_client_data(device, &ipoib_client);
>>>> diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
>>>> index 65994a1..2bf9094 100644
>>>> --- a/include/rdma/ib_verbs.h
>>>> +++ b/include/rdma/ib_verbs.h
>>>> @@ -1743,6 +1743,30 @@ int ib_query_port(struct ib_device *device,
>>>>  enum rdma_link_layer rdma_port_get_link_layer(struct ib_device *device,
>>>>                             u8 port_num);
>>>>  
>>>> +static inline int rdma_transport_is_ib(struct ib_device *device)
>>>> +{
>>>> +    return rdma_node_get_transport(device->node_type)
>>>> +            == RDMA_TRANSPORT_IB;
>>>> +}
>>>> +
>>>> +static inline int rdma_transport_is_iwarp(struct ib_device *device)
>>>> +{
>>>> +    return rdma_node_get_transport(device->node_type)
>>>> +            == RDMA_TRANSPORT_IWARP;
>>>> +}
>>>> +
>>>> +static inline int rdma_port_ll_is_ib(struct ib_device *device, u8 port_num)
>>>> +{
>>>> +    return rdma_port_get_link_layer(device, port_num)
>>>> +            == IB_LINK_LAYER_INFINIBAND;
>>>> +}
>>>> +
>>>> +static inline int rdma_port_ll_is_eth(struct ib_device *device, u8 port_num)
>>>> +{
>>>> +    return rdma_port_get_link_layer(device, port_num)
>>>> +            == IB_LINK_LAYER_ETHERNET;
>>>> +}
>>>> +
>>>>  int ib_query_gid(struct ib_device *device,
>>>>           u8 port_num, int index, union ib_gid *gid);
>>>>  
>>>> diff --git a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
>>>> index e011027..a7b5891 100644
>>>> --- a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
>>>> +++ b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
>>>> @@ -118,8 +118,7 @@ static void rdma_build_arg_xdr(struct svc_rqst *rqstp,
>>>>  
>>>>  static int rdma_read_max_sge(struct svcxprt_rdma *xprt, int sge_count)
>>>>  {
>>>> -    if (rdma_node_get_transport(xprt->sc_cm_id->device->node_type) ==
>>>> -         RDMA_TRANSPORT_IWARP)
>>>> +    if (rdma_transport_is_iwarp(xprt->sc_cm_id->device))
>>>>          return 1;
>>>>      else
>>>>          return min_t(int, sge_count, xprt->sc_max_sge);
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
>> the body of a message to majordomo@...r.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ