[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1427732576.21101.205.camel@redhat.com>
Date: Mon, 30 Mar 2015 12:22:56 -0400
From: Doug Ledford <dledford@...hat.com>
To: Michael Wang <yun.wang@...fitbricks.com>
Cc: Roland Dreier <roland@...nel.org>,
Sean Hefty <sean.hefty@...el.com>,
Hal Rosenstock <hal.rosenstock@...il.com>,
Ira Weiny <ira.weiny@...el.com>, linux-rdma@...r.kernel.org,
linux-kernel@...r.kernel.org, linux-nfs@...r.kernel.org,
netdev@...r.kernel.org, "J. Bruce Fields" <bfields@...ldses.org>,
Trond Myklebust <trond.myklebust@...marydata.com>,
"David S. Miller" <davem@...emloft.net>,
Or Gerlitz <ogerlitz@...lanox.com>,
Moni Shoua <monis@...lanox.com>,
PJ Waskiewicz <pj.waskiewicz@...idfire.com>,
Tatyana Nikolova <Tatyana.E.Nikolova@...el.com>,
Yan Burman <yanb@...lanox.com>,
Jack Morgenstein <jackm@....mellanox.co.il>,
Bart Van Assche <bvanassche@....org>,
Yann Droneaud <ydroneaud@...eya.com>,
Colin Ian King <colin.king@...onical.com>,
Majd Dibbiny <majd@...lanox.com>,
Jiri Kosina <jkosina@...e.cz>,
Matan Barak <matanb@...lanox.com>,
Alex Estrin <alex.estrin@...el.com>,
Eric Dumazet <edumazet@...gle.com>,
Erez Shitrit <erezsh@...lanox.com>,
Sagi Grimberg <sagig@...lanox.com>,
Haggai Eran <haggaie@...lanox.com>,
Shachar Raindel <raindel@...lanox.com>,
Mike Marciniszyn <mike.marciniszyn@...el.com>,
Steve Wise <swise@...ngridcomputing.com>,
Tom Tucker <tom@....us>, Chuck Lever <chuck.lever@...cle.com>
Subject: Re: [PATCH 01/11] IB/Verbs: Use helpers to check transport and link
layer
On Mon, 2015-03-30 at 18:14 +0200, Michael Wang wrote:
> Hi, Doug
>
> Thanks for the comments :-)
>
> On 03/30/2015 05:56 PM, Doug Ledford wrote:
> > On Fri, 2015-03-27 at 16:40 +0100, Michael Wang wrote:
> >> We have so much places to check transport type and link layer type, it's now
> >> make sense to introduce some helpers in order to refine the lengthy code.
> >>
> >> This patch will introduce helpers:
> >> rdma_transport_is_ib()
> >> rdma_transport_is_iwarp()
> >> rdma_port_ll_is_ib()
> >> rdma_port_ll_is_eth()
> >> and use them to save some code for us.
> > If the end result is to do something like I proposed, then why take this
> > intermediate step that just has to be backed out later?
>
> The problem is that I found there are still many places our new
> mechanism may could not take care, especially inside device driver,
> this is just try to collect the issues together as a basement so we can
> gradually eliminate them.
There is no "gradually eliminate them" to the suggestion I made.
Remember, my suggestion was to remove the transport and link_layer items
from the port settings and replace it with just one transport item that
is a bitmask of the possible transport types. This can not be done
gradually, it must be a complete change all at once as the two methods
of setting things are incompatible. As there is only one out of tree
driver that I know of, lustre, we can give them the information they
need to make their driver work both before and after the change.
> Sure if finally we do capture all the cases, we can just get rid of
> this one, but I guess it won't be that easy to directly jump into
> next stage :-P
>
> As I could imaging, after this reform, next stage could be introducing
> the new mechanism without changing device driver, and the last
> stage is to asking vendor adapt their code into the new mechanism.
>
> >
> > In other words, if our end goal is to have
> >
> > rdma_transport_is_ib()
> > rdma_transport_is_iwarp()
> > rdma_transport_is_roce()
> > rdma_transport_is_opa()
> >
> > Then we should skip doing rdma_port_ll_is_*() as the answers to these
> > items would be implied by rdma_transport_is_roce() and such.
>
> Great if we achieved that ;-) but currently I just wondering maybe
> these helpers can only cover part of the cases where we check
> transport and link layer, there are still some cases we'll need the
> very rough helper to save some code and make things clean~
>
> Regards,
> Michael Wang
>
>
> >
> >> Cc: Jason Gunthorpe <jgunthorpe@...idianresearch.com>
> >> Cc: Doug Ledford <dledford@...hat.com>
> >> Cc: Ira Weiny <ira.weiny@...el.com>
> >> Cc: Sean Hefty <sean.hefty@...el.com>
> >> Signed-off-by: Michael Wang <yun.wang@...fitbricks.com>
> >> ---
> >> drivers/infiniband/core/agent.c | 2 +-
> >> drivers/infiniband/core/cm.c | 2 +-
> >> drivers/infiniband/core/cma.c | 27 ++++++++++++---------------
> >> drivers/infiniband/core/mad.c | 6 +++---
> >> drivers/infiniband/core/multicast.c | 11 ++++-------
> >> drivers/infiniband/core/sa_query.c | 14 +++++++-------
> >> drivers/infiniband/core/ucm.c | 3 +--
> >> drivers/infiniband/core/user_mad.c | 2 +-
> >> drivers/infiniband/core/verbs.c | 5 ++---
> >> drivers/infiniband/hw/mlx4/ah.c | 2 +-
> >> drivers/infiniband/hw/mlx4/cq.c | 4 +---
> >> drivers/infiniband/hw/mlx4/mad.c | 14 ++++----------
> >> drivers/infiniband/hw/mlx4/main.c | 8 +++-----
> >> drivers/infiniband/hw/mlx4/mlx4_ib.h | 2 +-
> >> drivers/infiniband/hw/mlx4/qp.c | 21 +++++++--------------
> >> drivers/infiniband/hw/mlx4/sysfs.c | 6 ++----
> >> drivers/infiniband/ulp/ipoib/ipoib_main.c | 6 +++---
> >> include/rdma/ib_verbs.h | 24 ++++++++++++++++++++++++
> >> net/sunrpc/xprtrdma/svc_rdma_recvfrom.c | 3 +--
> >> 19 files changed, 79 insertions(+), 83 deletions(-)
> >>
> >> diff --git a/drivers/infiniband/core/agent.c b/drivers/infiniband/core/agent.c
> >> index f6d2961..27f1bec 100644
> >> --- a/drivers/infiniband/core/agent.c
> >> +++ b/drivers/infiniband/core/agent.c
> >> @@ -156,7 +156,7 @@ int ib_agent_port_open(struct ib_device *device, int port_num)
> >> goto error1;
> >> }
> >>
> >> - if (rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_INFINIBAND) {
> >> + if (rdma_port_ll_is_ib(device, port_num)) {
> >> /* Obtain send only MAD agent for SMI QP */
> >> port_priv->agent[0] = ib_register_mad_agent(device, port_num,
> >> IB_QPT_SMI, NULL, 0,
> >> diff --git a/drivers/infiniband/core/cm.c b/drivers/infiniband/core/cm.c
> >> index e28a494..2c72e9e 100644
> >> --- a/drivers/infiniband/core/cm.c
> >> +++ b/drivers/infiniband/core/cm.c
> >> @@ -3762,7 +3762,7 @@ static void cm_add_one(struct ib_device *ib_device)
> >> int ret;
> >> u8 i;
> >>
> >> - if (rdma_node_get_transport(ib_device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(ib_device))
> >> return;
> >>
> >> cm_dev = kzalloc(sizeof(*cm_dev) + sizeof(*port) *
> >> diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c
> >> index d570030..668e955 100644
> >> --- a/drivers/infiniband/core/cma.c
> >> +++ b/drivers/infiniband/core/cma.c
> >> @@ -375,8 +375,8 @@ static int cma_acquire_dev(struct rdma_id_private *id_priv,
> >> listen_id_priv->id.port_num) == dev_ll) {
> >> cma_dev = listen_id_priv->cma_dev;
> >> port = listen_id_priv->id.port_num;
> >> - if (rdma_node_get_transport(cma_dev->device->node_type) == RDMA_TRANSPORT_IB &&
> >> - rdma_port_get_link_layer(cma_dev->device, port) == IB_LINK_LAYER_ETHERNET)
> >> + if (rdma_transport_is_ib(cma_dev->device) &&
> >> + rdma_port_ll_is_eth(cma_dev->device, port))
> >> ret = ib_find_cached_gid(cma_dev->device, &iboe_gid,
> >> &found_port, NULL);
> >> else
> >> @@ -395,8 +395,8 @@ static int cma_acquire_dev(struct rdma_id_private *id_priv,
> >> listen_id_priv->id.port_num == port)
> >> continue;
> >> if (rdma_port_get_link_layer(cma_dev->device, port) == dev_ll) {
> >> - if (rdma_node_get_transport(cma_dev->device->node_type) == RDMA_TRANSPORT_IB &&
> >> - rdma_port_get_link_layer(cma_dev->device, port) == IB_LINK_LAYER_ETHERNET)
> >> + if (rdma_transport_is_ib(cma_dev->device) &&
> >> + rdma_port_ll_is_eth(cma_dev->device, port))
> >> ret = ib_find_cached_gid(cma_dev->device, &iboe_gid, &found_port, NULL);
> >> else
> >> ret = ib_find_cached_gid(cma_dev->device, &gid, &found_port, NULL);
> >> @@ -435,7 +435,7 @@ static int cma_resolve_ib_dev(struct rdma_id_private *id_priv)
> >> pkey = ntohs(addr->sib_pkey);
> >>
> >> list_for_each_entry(cur_dev, &dev_list, list) {
> >> - if (rdma_node_get_transport(cur_dev->device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(cur_dev->device))
> >> continue;
> >>
> >> for (p = 1; p <= cur_dev->device->phys_port_cnt; ++p) {
> >> @@ -633,10 +633,8 @@ static int cma_modify_qp_rtr(struct rdma_id_private *id_priv,
> >> if (ret)
> >> goto out;
> >>
> >> - if (rdma_node_get_transport(id_priv->cma_dev->device->node_type)
> >> - == RDMA_TRANSPORT_IB &&
> >> - rdma_port_get_link_layer(id_priv->id.device, id_priv->id.port_num)
> >> - == IB_LINK_LAYER_ETHERNET) {
> >> + if (rdma_transport_is_ib(id_priv->cma_dev->device) &&
> >> + rdma_port_ll_is_eth(id_priv->id.device, id_priv->id.port_num)) {
> >> ret = rdma_addr_find_smac_by_sgid(&sgid, qp_attr.smac, NULL);
> >>
> >> if (ret)
> >> @@ -700,8 +698,7 @@ static int cma_ib_init_qp_attr(struct rdma_id_private *id_priv,
> >> int ret;
> >> u16 pkey;
> >>
> >> - if (rdma_port_get_link_layer(id_priv->id.device, id_priv->id.port_num) ==
> >> - IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(id_priv->id.device, id_priv->id.port_num))
> >> pkey = ib_addr_get_pkey(dev_addr);
> >> else
> >> pkey = 0xffff;
> >> @@ -1626,7 +1623,7 @@ static void cma_listen_on_dev(struct rdma_id_private *id_priv,
> >> int ret;
> >>
> >> if (cma_family(id_priv) == AF_IB &&
> >> - rdma_node_get_transport(cma_dev->device->node_type) != RDMA_TRANSPORT_IB)
> >> + !rdma_transport_is_ib(cma_dev->device))
> >> return;
> >>
> >> id = rdma_create_id(cma_listen_handler, id_priv, id_priv->id.ps,
> >> @@ -2028,7 +2025,7 @@ static int cma_bind_loopback(struct rdma_id_private *id_priv)
> >> mutex_lock(&lock);
> >> list_for_each_entry(cur_dev, &dev_list, list) {
> >> if (cma_family(id_priv) == AF_IB &&
> >> - rdma_node_get_transport(cur_dev->device->node_type) != RDMA_TRANSPORT_IB)
> >> + !rdma_transport_is_ib(cur_dev->device))
> >> continue;
> >>
> >> if (!cma_dev)
> >> @@ -2060,7 +2057,7 @@ port_found:
> >> goto out;
> >>
> >> id_priv->id.route.addr.dev_addr.dev_type =
> >> - (rdma_port_get_link_layer(cma_dev->device, p) == IB_LINK_LAYER_INFINIBAND) ?
> >> + (rdma_port_ll_is_ib(cma_dev->device, p)) ?
> >> ARPHRD_INFINIBAND : ARPHRD_ETHER;
> >>
> >> rdma_addr_set_sgid(&id_priv->id.route.addr.dev_addr, &gid);
> >> @@ -3405,7 +3402,7 @@ void rdma_leave_multicast(struct rdma_cm_id *id, struct sockaddr *addr)
> >> ib_detach_mcast(id->qp,
> >> &mc->multicast.ib->rec.mgid,
> >> be16_to_cpu(mc->multicast.ib->rec.mlid));
> >> - if (rdma_node_get_transport(id_priv->cma_dev->device->node_type) == RDMA_TRANSPORT_IB) {
> >> + if (rdma_transport_is_ib(id_priv->cma_dev->device)) {
> >> switch (rdma_port_get_link_layer(id->device, id->port_num)) {
> >> case IB_LINK_LAYER_INFINIBAND:
> >> ib_sa_free_multicast(mc->multicast.ib);
> >> diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c
> >> index 74c30f4..23cf9e8 100644
> >> --- a/drivers/infiniband/core/mad.c
> >> +++ b/drivers/infiniband/core/mad.c
> >> @@ -2938,7 +2938,7 @@ static int ib_mad_port_open(struct ib_device *device,
> >> init_mad_qp(port_priv, &port_priv->qp_info[1]);
> >>
> >> cq_size = mad_sendq_size + mad_recvq_size;
> >> - has_smi = rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_INFINIBAND;
> >> + has_smi = rdma_port_ll_is_ib(device, port_num);
> >> if (has_smi)
> >> cq_size *= 2;
> >>
> >> @@ -3057,7 +3057,7 @@ static void ib_mad_init_device(struct ib_device *device)
> >> {
> >> int start, end, i;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> if (device->node_type == RDMA_NODE_IB_SWITCH) {
> >> @@ -3102,7 +3102,7 @@ static void ib_mad_remove_device(struct ib_device *device)
> >> {
> >> int i, num_ports, cur_port;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> if (device->node_type == RDMA_NODE_IB_SWITCH) {
> >> diff --git a/drivers/infiniband/core/multicast.c b/drivers/infiniband/core/multicast.c
> >> index fa17b55..17573ff 100644
> >> --- a/drivers/infiniband/core/multicast.c
> >> +++ b/drivers/infiniband/core/multicast.c
> >> @@ -780,8 +780,7 @@ static void mcast_event_handler(struct ib_event_handler *handler,
> >> int index;
> >>
> >> dev = container_of(handler, struct mcast_device, event_handler);
> >> - if (rdma_port_get_link_layer(dev->device, event->element.port_num) !=
> >> - IB_LINK_LAYER_INFINIBAND)
> >> + if (!rdma_port_ll_is_ib(dev->device, event->element.port_num))
> >> return;
> >>
> >> index = event->element.port_num - dev->start_port;
> >> @@ -808,7 +807,7 @@ static void mcast_add_one(struct ib_device *device)
> >> int i;
> >> int count = 0;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> dev = kmalloc(sizeof *dev + device->phys_port_cnt * sizeof *port,
> >> @@ -824,8 +823,7 @@ static void mcast_add_one(struct ib_device *device)
> >> }
> >>
> >> for (i = 0; i <= dev->end_port - dev->start_port; i++) {
> >> - if (rdma_port_get_link_layer(device, dev->start_port + i) !=
> >> - IB_LINK_LAYER_INFINIBAND)
> >> + if (!rdma_port_ll_is_ib(device, dev->start_port + i))
> >> continue;
> >> port = &dev->port[i];
> >> port->dev = dev;
> >> @@ -863,8 +861,7 @@ static void mcast_remove_one(struct ib_device *device)
> >> flush_workqueue(mcast_wq);
> >>
> >> for (i = 0; i <= dev->end_port - dev->start_port; i++) {
> >> - if (rdma_port_get_link_layer(device, dev->start_port + i) ==
> >> - IB_LINK_LAYER_INFINIBAND) {
> >> + if (rdma_port_ll_is_ib(device, dev->start_port + i)) {
> >> port = &dev->port[i];
> >> deref_port(port);
> >> wait_for_completion(&port->comp);
> >> diff --git a/drivers/infiniband/core/sa_query.c b/drivers/infiniband/core/sa_query.c
> >> index c38f030..d95d25f 100644
> >> --- a/drivers/infiniband/core/sa_query.c
> >> +++ b/drivers/infiniband/core/sa_query.c
> >> @@ -450,7 +450,7 @@ static void ib_sa_event(struct ib_event_handler *handler, struct ib_event *event
> >> struct ib_sa_port *port =
> >> &sa_dev->port[event->element.port_num - sa_dev->start_port];
> >>
> >> - if (rdma_port_get_link_layer(handler->device, port->port_num) != IB_LINK_LAYER_INFINIBAND)
> >> + if (!rdma_port_ll_is_ib(handler->device, port->port_num))
> >> return;
> >>
> >> spin_lock_irqsave(&port->ah_lock, flags);
> >> @@ -540,7 +540,7 @@ int ib_init_ah_from_path(struct ib_device *device, u8 port_num,
> >> ah_attr->port_num = port_num;
> >> ah_attr->static_rate = rec->rate;
> >>
> >> - force_grh = rdma_port_get_link_layer(device, port_num) == IB_LINK_LAYER_ETHERNET;
> >> + force_grh = rdma_port_ll_is_eth(device, port_num);
> >>
> >> if (rec->hop_limit > 1 || force_grh) {
> >> ah_attr->ah_flags = IB_AH_GRH;
> >> @@ -1154,7 +1154,7 @@ static void ib_sa_add_one(struct ib_device *device)
> >> struct ib_sa_device *sa_dev;
> >> int s, e, i;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> if (device->node_type == RDMA_NODE_IB_SWITCH)
> >> @@ -1175,7 +1175,7 @@ static void ib_sa_add_one(struct ib_device *device)
> >>
> >> for (i = 0; i <= e - s; ++i) {
> >> spin_lock_init(&sa_dev->port[i].ah_lock);
> >> - if (rdma_port_get_link_layer(device, i + 1) != IB_LINK_LAYER_INFINIBAND)
> >> + if (!rdma_port_ll_is_ib(device, i + 1))
> >> continue;
> >>
> >> sa_dev->port[i].sm_ah = NULL;
> >> @@ -1205,14 +1205,14 @@ static void ib_sa_add_one(struct ib_device *device)
> >> goto err;
> >>
> >> for (i = 0; i <= e - s; ++i)
> >> - if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(device, i + 1))
> >> update_sm_ah(&sa_dev->port[i].update_task);
> >>
> >> return;
> >>
> >> err:
> >> while (--i >= 0)
> >> - if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(device, i + 1))
> >> ib_unregister_mad_agent(sa_dev->port[i].agent);
> >>
> >> kfree(sa_dev);
> >> @@ -1233,7 +1233,7 @@ static void ib_sa_remove_one(struct ib_device *device)
> >> flush_workqueue(ib_wq);
> >>
> >> for (i = 0; i <= sa_dev->end_port - sa_dev->start_port; ++i) {
> >> - if (rdma_port_get_link_layer(device, i + 1) == IB_LINK_LAYER_INFINIBAND) {
> >> + if (rdma_port_ll_is_ib(device, i + 1)) {
> >> ib_unregister_mad_agent(sa_dev->port[i].agent);
> >> if (sa_dev->port[i].sm_ah)
> >> kref_put(&sa_dev->port[i].sm_ah->ref, free_sm_ah);
> >> diff --git a/drivers/infiniband/core/ucm.c b/drivers/infiniband/core/ucm.c
> >> index f2f6393..ddbe0b4 100644
> >> --- a/drivers/infiniband/core/ucm.c
> >> +++ b/drivers/infiniband/core/ucm.c
> >> @@ -1253,8 +1253,7 @@ static void ib_ucm_add_one(struct ib_device *device)
> >> dev_t base;
> >> struct ib_ucm_device *ucm_dev;
> >>
> >> - if (!device->alloc_ucontext ||
> >> - rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!device->alloc_ucontext || !rdma_transport_is_ib(device))
> >> return;
> >>
> >> ucm_dev = kzalloc(sizeof *ucm_dev, GFP_KERNEL);
> >> diff --git a/drivers/infiniband/core/user_mad.c b/drivers/infiniband/core/user_mad.c
> >> index 928cdd2..28a8b30 100644
> >> --- a/drivers/infiniband/core/user_mad.c
> >> +++ b/drivers/infiniband/core/user_mad.c
> >> @@ -1274,7 +1274,7 @@ static void ib_umad_add_one(struct ib_device *device)
> >> struct ib_umad_device *umad_dev;
> >> int s, e, i;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> if (device->node_type == RDMA_NODE_IB_SWITCH)
> >> diff --git a/drivers/infiniband/core/verbs.c b/drivers/infiniband/core/verbs.c
> >> index f93eb8d..d8d015a 100644
> >> --- a/drivers/infiniband/core/verbs.c
> >> +++ b/drivers/infiniband/core/verbs.c
> >> @@ -198,8 +198,7 @@ int ib_init_ah_from_wc(struct ib_device *device, u8 port_num, struct ib_wc *wc,
> >> u32 flow_class;
> >> u16 gid_index;
> >> int ret;
> >> - int is_eth = (rdma_port_get_link_layer(device, port_num) ==
> >> - IB_LINK_LAYER_ETHERNET);
> >> + int is_eth = (rdma_port_ll_is_eth(device, port_num));
> >>
> >> memset(ah_attr, 0, sizeof *ah_attr);
> >> if (is_eth) {
> >> @@ -871,7 +870,7 @@ int ib_resolve_eth_l2_attrs(struct ib_qp *qp,
> >> union ib_gid sgid;
> >>
> >> if ((*qp_attr_mask & IB_QP_AV) &&
> >> - (rdma_port_get_link_layer(qp->device, qp_attr->ah_attr.port_num) == IB_LINK_LAYER_ETHERNET)) {
> >> + (rdma_port_ll_is_eth(qp->device, qp_attr->ah_attr.port_num))) {
> >> ret = ib_query_gid(qp->device, qp_attr->ah_attr.port_num,
> >> qp_attr->ah_attr.grh.sgid_index, &sgid);
> >> if (ret)
> >> diff --git a/drivers/infiniband/hw/mlx4/ah.c b/drivers/infiniband/hw/mlx4/ah.c
> >> index 2d8c339..829eb60 100644
> >> --- a/drivers/infiniband/hw/mlx4/ah.c
> >> +++ b/drivers/infiniband/hw/mlx4/ah.c
> >> @@ -118,7 +118,7 @@ struct ib_ah *mlx4_ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
> >> if (!ah)
> >> return ERR_PTR(-ENOMEM);
> >>
> >> - if (rdma_port_get_link_layer(pd->device, ah_attr->port_num) == IB_LINK_LAYER_ETHERNET) {
> >> + if (rdma_port_ll_is_eth(pd->device, ah_attr->port_num)) {
> >> if (!(ah_attr->ah_flags & IB_AH_GRH)) {
> >> ret = ERR_PTR(-EINVAL);
> >> } else {
> >> diff --git a/drivers/infiniband/hw/mlx4/cq.c b/drivers/infiniband/hw/mlx4/cq.c
> >> index cb63ecd..0417f03 100644
> >> --- a/drivers/infiniband/hw/mlx4/cq.c
> >> +++ b/drivers/infiniband/hw/mlx4/cq.c
> >> @@ -789,9 +789,7 @@ repoll:
> >> break;
> >> }
> >>
> >> - is_eth = (rdma_port_get_link_layer(wc->qp->device,
> >> - (*cur_qp)->port) ==
> >> - IB_LINK_LAYER_ETHERNET);
> >> + is_eth = (rdma_port_ll_is_eth(wc->qp->device, (*cur_qp)->port));
> >> if (mlx4_is_mfunc(to_mdev(cq->ibcq.device)->dev)) {
> >> if ((*cur_qp)->mlx4_ib_qp_type &
> >> (MLX4_IB_QPT_PROXY_SMI_OWNER |
> >> diff --git a/drivers/infiniband/hw/mlx4/mad.c b/drivers/infiniband/hw/mlx4/mad.c
> >> index 82a7dd8..4736fc7 100644
> >> --- a/drivers/infiniband/hw/mlx4/mad.c
> >> +++ b/drivers/infiniband/hw/mlx4/mad.c
> >> @@ -606,12 +606,7 @@ static int mlx4_ib_demux_mad(struct ib_device *ibdev, u8 port,
> >> int err;
> >> int slave;
> >> u8 *slave_id;
> >> - int is_eth = 0;
> >> -
> >> - if (rdma_port_get_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND)
> >> - is_eth = 0;
> >> - else
> >> - is_eth = 1;
> >> + int is_eth = rdma_port_ll_is_eth(ibdev, port);
> >>
> >> if (is_eth) {
> >> if (!(wc->wc_flags & IB_WC_GRH)) {
> >> @@ -1252,7 +1247,7 @@ out:
> >>
> >> static int get_slave_base_gid_ix(struct mlx4_ib_dev *dev, int slave, int port)
> >> {
> >> - if (rdma_port_get_link_layer(&dev->ib_dev, port) == IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(&dev->ib_dev, port))
> >> return slave;
> >> return mlx4_get_base_gid_ix(dev->dev, slave, port);
> >> }
> >> @@ -1260,7 +1255,7 @@ static int get_slave_base_gid_ix(struct mlx4_ib_dev *dev, int slave, int port)
> >> static void fill_in_real_sgid_index(struct mlx4_ib_dev *dev, int slave, int port,
> >> struct ib_ah_attr *ah_attr)
> >> {
> >> - if (rdma_port_get_link_layer(&dev->ib_dev, port) == IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(&dev->ib_dev, port))
> >> ah_attr->grh.sgid_index = slave;
> >> else
> >> ah_attr->grh.sgid_index += get_slave_base_gid_ix(dev, slave, port);
> >> @@ -1758,8 +1753,7 @@ static int create_pv_resources(struct ib_device *ibdev, int slave, int port,
> >>
> >> ctx->state = DEMUX_PV_STATE_STARTING;
> >> /* have QP0 only if link layer is IB */
> >> - if (rdma_port_get_link_layer(ibdev, ctx->port) ==
> >> - IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(ibdev, ctx->port))
> >> ctx->has_smi = 1;
> >>
> >> if (ctx->has_smi) {
> >> diff --git a/drivers/infiniband/hw/mlx4/main.c b/drivers/infiniband/hw/mlx4/main.c
> >> index 0b280b1..f445f4c 100644
> >> --- a/drivers/infiniband/hw/mlx4/main.c
> >> +++ b/drivers/infiniband/hw/mlx4/main.c
> >> @@ -482,7 +482,7 @@ static int iboe_query_gid(struct ib_device *ibdev, u8 port, int index,
> >> static int mlx4_ib_query_gid(struct ib_device *ibdev, u8 port, int index,
> >> union ib_gid *gid)
> >> {
> >> - if (rdma_port_get_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND)
> >> + if (rdma_port_ll_is_ib(ibdev, port))
> >> return __mlx4_ib_query_gid(ibdev, port, index, gid, 0);
> >> else
> >> return iboe_query_gid(ibdev, port, index, gid);
> >> @@ -1801,8 +1801,7 @@ static int mlx4_ib_init_gid_table(struct mlx4_ib_dev *ibdev)
> >> int err = 0;
> >>
> >> for (i = 1; i <= ibdev->num_ports; ++i) {
> >> - if (rdma_port_get_link_layer(&ibdev->ib_dev, i) ==
> >> - IB_LINK_LAYER_ETHERNET) {
> >> + if (rdma_port_ll_is_eth(&ibdev->ib_dev, i)) {
> >> err = reset_gid_table(ibdev, i);
> >> if (err)
> >> goto out;
> >> @@ -2554,8 +2553,7 @@ static void mlx4_ib_event(struct mlx4_dev *dev, void *ibdev_ptr,
> >> if (p > ibdev->num_ports)
> >> return;
> >> if (mlx4_is_master(dev) &&
> >> - rdma_port_get_link_layer(&ibdev->ib_dev, p) ==
> >> - IB_LINK_LAYER_INFINIBAND) {
> >> + rdma_port_ll_is_ib(&ibdev->ib_dev, p)) {
> >> mlx4_ib_invalidate_all_guid_record(ibdev, p);
> >> }
> >> ibev.event = IB_EVENT_PORT_ACTIVE;
> >> diff --git a/drivers/infiniband/hw/mlx4/mlx4_ib.h b/drivers/infiniband/hw/mlx4/mlx4_ib.h
> >> index 6eb743f..1befeb8 100644
> >> --- a/drivers/infiniband/hw/mlx4/mlx4_ib.h
> >> +++ b/drivers/infiniband/hw/mlx4/mlx4_ib.h
> >> @@ -712,7 +712,7 @@ static inline bool mlx4_ib_ah_grh_present(struct mlx4_ib_ah *ah)
> >> {
> >> u8 port = be32_to_cpu(ah->av.ib.port_pd) >> 24 & 3;
> >>
> >> - if (rdma_port_get_link_layer(ah->ibah.device, port) == IB_LINK_LAYER_ETHERNET)
> >> + if (rdma_port_ll_is_eth(ah->ibah.device, port))
> >> return true;
> >>
> >> return !!(ah->av.ib.g_slid & 0x80);
> >> diff --git a/drivers/infiniband/hw/mlx4/qp.c b/drivers/infiniband/hw/mlx4/qp.c
> >> index c880329..bd2f557 100644
> >> --- a/drivers/infiniband/hw/mlx4/qp.c
> >> +++ b/drivers/infiniband/hw/mlx4/qp.c
> >> @@ -1248,8 +1248,7 @@ static int _mlx4_set_path(struct mlx4_ib_dev *dev, const struct ib_ah_attr *ah,
> >> u64 smac, u16 vlan_tag, struct mlx4_qp_path *path,
> >> struct mlx4_roce_smac_vlan_info *smac_info, u8 port)
> >> {
> >> - int is_eth = rdma_port_get_link_layer(&dev->ib_dev, port) ==
> >> - IB_LINK_LAYER_ETHERNET;
> >> + int is_eth = rdma_port_ll_is_eth(&dev->ib_dev, port);
> >> int vidx;
> >> int smac_index;
> >> int err;
> >> @@ -1433,8 +1432,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
> >>
> >> /* APM is not supported under RoCE */
> >> if (attr_mask & IB_QP_ALT_PATH &&
> >> - rdma_port_get_link_layer(&dev->ib_dev, qp->port) ==
> >> - IB_LINK_LAYER_ETHERNET)
> >> + rdma_port_ll_is_eth(&dev->ib_dev, qp->port))
> >> return -ENOTSUPP;
> >>
> >> context = kzalloc(sizeof *context, GFP_KERNEL);
> >> @@ -1664,8 +1662,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
> >> context->pri_path.fl = 0x80;
> >> context->pri_path.sched_queue |= MLX4_IB_DEFAULT_SCHED_QUEUE;
> >> }
> >> - if (rdma_port_get_link_layer(&dev->ib_dev, qp->port) ==
> >> - IB_LINK_LAYER_ETHERNET) {
> >> + if (rdma_port_ll_is_eth(&dev->ib_dev, qp->port)) {
> >> if (qp->mlx4_ib_qp_type == MLX4_IB_QPT_TUN_GSI ||
> >> qp->mlx4_ib_qp_type == MLX4_IB_QPT_GSI)
> >> context->pri_path.feup = 1 << 7; /* don't fsm */
> >> @@ -1695,9 +1692,7 @@ static int __mlx4_ib_modify_qp(struct ib_qp *ibqp,
> >> }
> >>
> >> if (ibqp->qp_type == IB_QPT_UD && (new_state == IB_QPS_RTR)) {
> >> - int is_eth = rdma_port_get_link_layer(
> >> - &dev->ib_dev, qp->port) ==
> >> - IB_LINK_LAYER_ETHERNET;
> >> + int is_eth = rdma_port_ll_is_eth(&dev->ib_dev, qp->port);
> >> if (is_eth) {
> >> context->pri_path.ackto = MLX4_IB_LINK_TYPE_ETH;
> >> optpar |= MLX4_QP_OPTPAR_PRIMARY_ADDR_PATH;
> >> @@ -1927,8 +1922,7 @@ int mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
> >> }
> >>
> >> if ((attr_mask & IB_QP_PORT) && (ibqp->qp_type == IB_QPT_RAW_PACKET) &&
> >> - (rdma_port_get_link_layer(&dev->ib_dev, attr->port_num) !=
> >> - IB_LINK_LAYER_ETHERNET))
> >> + !rdma_port_ll_is_eth(&dev->ib_dev, attr->port_num))
> >> goto out;
> >>
> >> if (attr_mask & IB_QP_PKEY_INDEX) {
> >> @@ -2132,7 +2126,7 @@ static int build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr,
> >> for (i = 0; i < wr->num_sge; ++i)
> >> send_size += wr->sg_list[i].length;
> >>
> >> - is_eth = rdma_port_get_link_layer(sqp->qp.ibqp.device, sqp->qp.port) == IB_LINK_LAYER_ETHERNET;
> >> + is_eth = rdma_port_ll_is_eth(sqp->qp.ibqp.device, sqp->qp.port);
> >> is_grh = mlx4_ib_ah_grh_present(ah);
> >> if (is_eth) {
> >> if (mlx4_is_mfunc(to_mdev(ib_dev)->dev)) {
> >> @@ -3029,8 +3023,7 @@ static void to_ib_ah_attr(struct mlx4_ib_dev *ibdev, struct ib_ah_attr *ib_ah_at
> >> if (ib_ah_attr->port_num == 0 || ib_ah_attr->port_num > dev->caps.num_ports)
> >> return;
> >>
> >> - is_eth = rdma_port_get_link_layer(&ibdev->ib_dev, ib_ah_attr->port_num) ==
> >> - IB_LINK_LAYER_ETHERNET;
> >> + is_eth = rdma_port_ll_is_eth(&ibdev->ib_dev, ib_ah_attr->port_num);
> >> if (is_eth)
> >> ib_ah_attr->sl = ((path->sched_queue >> 3) & 0x7) |
> >> ((path->sched_queue & 4) << 1);
> >> diff --git a/drivers/infiniband/hw/mlx4/sysfs.c b/drivers/infiniband/hw/mlx4/sysfs.c
> >> index cb4c66e..d339b55 100644
> >> --- a/drivers/infiniband/hw/mlx4/sysfs.c
> >> +++ b/drivers/infiniband/hw/mlx4/sysfs.c
> >> @@ -610,8 +610,7 @@ static ssize_t sysfs_store_enable_smi_admin(struct device *dev,
> >>
> >> static int add_vf_smi_entries(struct mlx4_port *p)
> >> {
> >> - int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) ==
> >> - IB_LINK_LAYER_ETHERNET;
> >> + int is_eth = rdma_port_ll_is_eth(&p->dev->ib_dev, p->port_num);
> >> int ret;
> >>
> >> /* do not display entries if eth transport, or if master */
> >> @@ -645,8 +644,7 @@ static int add_vf_smi_entries(struct mlx4_port *p)
> >>
> >> static void remove_vf_smi_entries(struct mlx4_port *p)
> >> {
> >> - int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) ==
> >> - IB_LINK_LAYER_ETHERNET;
> >> + int is_eth = rdma_port_ll_is_eth(&p->dev->ib_dev, p->port_num);
> >>
> >> if (is_eth || p->slave == mlx4_master_func_num(p->dev->dev))
> >> return;
> >> diff --git a/drivers/infiniband/ulp/ipoib/ipoib_main.c b/drivers/infiniband/ulp/ipoib/ipoib_main.c
> >> index 58b5aa3..3341754 100644
> >> --- a/drivers/infiniband/ulp/ipoib/ipoib_main.c
> >> +++ b/drivers/infiniband/ulp/ipoib/ipoib_main.c
> >> @@ -1655,7 +1655,7 @@ static void ipoib_add_one(struct ib_device *device)
> >> struct ipoib_dev_priv *priv;
> >> int s, e, p;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> dev_list = kmalloc(sizeof *dev_list, GFP_KERNEL);
> >> @@ -1673,7 +1673,7 @@ static void ipoib_add_one(struct ib_device *device)
> >> }
> >>
> >> for (p = s; p <= e; ++p) {
> >> - if (rdma_port_get_link_layer(device, p) != IB_LINK_LAYER_INFINIBAND)
> >> + if (!rdma_port_ll_is_ib(device, p))
> >> continue;
> >> dev = ipoib_add_port("ib%d", device, p);
> >> if (!IS_ERR(dev)) {
> >> @@ -1690,7 +1690,7 @@ static void ipoib_remove_one(struct ib_device *device)
> >> struct ipoib_dev_priv *priv, *tmp;
> >> struct list_head *dev_list;
> >>
> >> - if (rdma_node_get_transport(device->node_type) != RDMA_TRANSPORT_IB)
> >> + if (!rdma_transport_is_ib(device))
> >> return;
> >>
> >> dev_list = ib_get_client_data(device, &ipoib_client);
> >> diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
> >> index 65994a1..2bf9094 100644
> >> --- a/include/rdma/ib_verbs.h
> >> +++ b/include/rdma/ib_verbs.h
> >> @@ -1743,6 +1743,30 @@ int ib_query_port(struct ib_device *device,
> >> enum rdma_link_layer rdma_port_get_link_layer(struct ib_device *device,
> >> u8 port_num);
> >>
> >> +static inline int rdma_transport_is_ib(struct ib_device *device)
> >> +{
> >> + return rdma_node_get_transport(device->node_type)
> >> + == RDMA_TRANSPORT_IB;
> >> +}
> >> +
> >> +static inline int rdma_transport_is_iwarp(struct ib_device *device)
> >> +{
> >> + return rdma_node_get_transport(device->node_type)
> >> + == RDMA_TRANSPORT_IWARP;
> >> +}
> >> +
> >> +static inline int rdma_port_ll_is_ib(struct ib_device *device, u8 port_num)
> >> +{
> >> + return rdma_port_get_link_layer(device, port_num)
> >> + == IB_LINK_LAYER_INFINIBAND;
> >> +}
> >> +
> >> +static inline int rdma_port_ll_is_eth(struct ib_device *device, u8 port_num)
> >> +{
> >> + return rdma_port_get_link_layer(device, port_num)
> >> + == IB_LINK_LAYER_ETHERNET;
> >> +}
> >> +
> >> int ib_query_gid(struct ib_device *device,
> >> u8 port_num, int index, union ib_gid *gid);
> >>
> >> diff --git a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
> >> index e011027..a7b5891 100644
> >> --- a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
> >> +++ b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
> >> @@ -118,8 +118,7 @@ static void rdma_build_arg_xdr(struct svc_rqst *rqstp,
> >>
> >> static int rdma_read_max_sge(struct svcxprt_rdma *xprt, int sge_count)
> >> {
> >> - if (rdma_node_get_transport(xprt->sc_cm_id->device->node_type) ==
> >> - RDMA_TRANSPORT_IWARP)
> >> + if (rdma_transport_is_iwarp(xprt->sc_cm_id->device))
> >> return 1;
> >> else
> >> return min_t(int, sge_count, xprt->sc_max_sge);
> >
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
--
Doug Ledford <dledford@...hat.com>
GPG KeyID: 0E572FDD
Download attachment "signature.asc" of type "application/pgp-signature" (820 bytes)
Powered by blists - more mailing lists