[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1537275826-27247-5-git-send-email-jan.dakinevich@virtuozzo.com>
Date: Tue, 18 Sep 2018 16:03:46 +0300
From: Jan Dakinevich <jan.dakinevich@...tuozzo.com>
To: Doug Ledford <dledford@...hat.com>, Jason Gunthorpe <jgg@...pe.ca>,
Yishai Hadas <yishaih@...lanox.com>,
Leon Romanovsky <leon@...nel.org>,
Parav Pandit <parav@...lanox.com>,
Mark Bloch <markb@...lanox.com>,
Daniel Jurgens <danielj@...lanox.com>,
Kees Cook <keescook@...omium.org>,
Kamal Heib <kamalheib1@...il.com>,
Bart Van Assche <bvanassche@....org>,
linux-rdma@...r.kernel.org, linux-kernel@...r.kernel.org
Cc: Denis Lunev <den@...tuozzo.com>,
Konstantin Khorenko <khorenko@...tuozzo.com>,
Jan Dakinevich <jan.dakinevich@...tuozzo.com>
Subject: [PATCH 4/4] IB/mlx4: move sriov field aside from mlx4_ib_dev
This is the 3rd patch of 3 of the work for decreasing size
of mlx4_ib_dev.
The field takes about 6K and could be safely allocated with kvzalloc.
Signed-off-by: Jan Dakinevich <jan.dakinevich@...tuozzo.com>
---
drivers/infiniband/hw/mlx4/alias_GUID.c | 192 ++++++++++++++++----------------
drivers/infiniband/hw/mlx4/cm.c | 32 +++---
drivers/infiniband/hw/mlx4/mad.c | 80 ++++++-------
drivers/infiniband/hw/mlx4/main.c | 17 ++-
drivers/infiniband/hw/mlx4/mcg.c | 4 +-
drivers/infiniband/hw/mlx4/mlx4_ib.h | 3 +-
drivers/infiniband/hw/mlx4/qp.c | 4 +-
drivers/infiniband/hw/mlx4/sysfs.c | 10 +-
8 files changed, 175 insertions(+), 167 deletions(-)
diff --git a/drivers/infiniband/hw/mlx4/alias_GUID.c b/drivers/infiniband/hw/mlx4/alias_GUID.c
index 155b4df..b5f794d 100644
--- a/drivers/infiniband/hw/mlx4/alias_GUID.c
+++ b/drivers/infiniband/hw/mlx4/alias_GUID.c
@@ -83,7 +83,7 @@ void mlx4_ib_update_cache_on_guid_change(struct mlx4_ib_dev *dev, int block_num,
if (!mlx4_is_master(dev->dev))
return;
- guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
+ guid_indexes = be64_to_cpu((__force __be64) dev->sriov->alias_guid.
ports_guid[port_num - 1].
all_rec_per_port[block_num].guid_indexes);
pr_debug("port: %d, guid_indexes: 0x%llx\n", port_num, guid_indexes);
@@ -99,7 +99,7 @@ void mlx4_ib_update_cache_on_guid_change(struct mlx4_ib_dev *dev, int block_num,
}
/* cache the guid: */
- memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id],
+ memcpy(&dev->sriov->demux[port_index].guid_cache[slave_id],
&p_data[i * GUID_REC_SIZE],
GUID_REC_SIZE);
} else
@@ -114,7 +114,7 @@ static __be64 get_cached_alias_guid(struct mlx4_ib_dev *dev, int port, int index
pr_err("%s: ERROR: asked for index:%d\n", __func__, index);
return (__force __be64) -1;
}
- return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index];
+ return *(__be64 *)&dev->sriov->demux[port - 1].guid_cache[index];
}
@@ -133,12 +133,12 @@ void mlx4_ib_slave_alias_guid_event(struct mlx4_ib_dev *dev, int slave,
unsigned long flags;
int do_work = 0;
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
- if (dev->sriov.alias_guid.ports_guid[port_index].state_flags &
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags);
+ if (dev->sriov->alias_guid.ports_guid[port_index].state_flags &
GUID_STATE_NEED_PORT_INIT)
goto unlock;
if (!slave_init) {
- curr_guid = *(__be64 *)&dev->sriov.
+ curr_guid = *(__be64 *)&dev->sriov->
alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].
all_recs[GUID_REC_SIZE * index];
@@ -151,24 +151,24 @@ void mlx4_ib_slave_alias_guid_event(struct mlx4_ib_dev *dev, int slave,
if (required_guid == cpu_to_be64(MLX4_GUID_FOR_DELETE_VAL))
goto unlock;
}
- *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index].
+ *(__be64 *)&dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].
all_recs[GUID_REC_SIZE * index] = required_guid;
- dev->sriov.alias_guid.ports_guid[port_index].
+ dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].guid_indexes
|= mlx4_ib_get_aguid_comp_mask_from_ix(index);
- dev->sriov.alias_guid.ports_guid[port_index].
+ dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].status
= MLX4_GUID_INFO_STATUS_IDLE;
/* set to run immediately */
- dev->sriov.alias_guid.ports_guid[port_index].
+ dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].time_to_run = 0;
- dev->sriov.alias_guid.ports_guid[port_index].
+ dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[record_num].
guids_retry_schedule[index] = 0;
do_work = 1;
unlock:
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags);
if (do_work)
mlx4_ib_init_alias_guid_work(dev, port_index);
@@ -201,9 +201,9 @@ void mlx4_ib_notify_slaves_on_guid_change(struct mlx4_ib_dev *dev,
if (!mlx4_is_master(dev->dev))
return;
- rec = &dev->sriov.alias_guid.ports_guid[port_num - 1].
+ rec = &dev->sriov->alias_guid.ports_guid[port_num - 1].
all_rec_per_port[block_num];
- guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
+ guid_indexes = be64_to_cpu((__force __be64) dev->sriov->alias_guid.
ports_guid[port_num - 1].
all_rec_per_port[block_num].guid_indexes);
pr_debug("port: %d, guid_indexes: 0x%llx\n", port_num, guid_indexes);
@@ -233,7 +233,7 @@ void mlx4_ib_notify_slaves_on_guid_change(struct mlx4_ib_dev *dev,
if (tmp_cur_ag != form_cache_ag)
continue;
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags);
required_value = *(__be64 *)&rec->all_recs[i * GUID_REC_SIZE];
if (required_value == cpu_to_be64(MLX4_GUID_FOR_DELETE_VAL))
@@ -245,12 +245,12 @@ void mlx4_ib_notify_slaves_on_guid_change(struct mlx4_ib_dev *dev,
} else {
/* may notify port down if value is 0 */
if (tmp_cur_ag != MLX4_NOT_SET_GUID) {
- spin_unlock_irqrestore(&dev->sriov.
+ spin_unlock_irqrestore(&dev->sriov->
alias_guid.ag_work_lock, flags);
continue;
}
}
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock,
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock,
flags);
mlx4_gen_guid_change_eqe(dev->dev, slave_id, port_num);
/*2 cases: Valid GUID, and Invalid Guid*/
@@ -304,7 +304,7 @@ static void aliasguid_query_handler(int status,
dev = cb_ctx->dev;
port_index = cb_ctx->port - 1;
- rec = &dev->sriov.alias_guid.ports_guid[port_index].
+ rec = &dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[cb_ctx->block_num];
if (status) {
@@ -324,10 +324,10 @@ static void aliasguid_query_handler(int status,
be16_to_cpu(guid_rec->lid), cb_ctx->port,
guid_rec->block_num);
- rec = &dev->sriov.alias_guid.ports_guid[port_index].
+ rec = &dev->sriov->alias_guid.ports_guid[port_index].
all_rec_per_port[guid_rec->block_num];
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags);
for (i = 0 ; i < NUM_ALIAS_GUID_IN_REC; i++) {
__be64 sm_response, required_val;
@@ -421,7 +421,7 @@ static void aliasguid_query_handler(int status,
} else {
rec->status = MLX4_GUID_INFO_STATUS_SET;
}
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags);
/*
The func is call here to close the cases when the
sm doesn't send smp, so in the sa response the driver
@@ -431,12 +431,12 @@ static void aliasguid_query_handler(int status,
cb_ctx->port,
guid_rec->guid_info_list);
out:
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
- if (!dev->sriov.is_going_down) {
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ if (!dev->sriov->is_going_down) {
get_low_record_time_index(dev, port_index, &resched_delay_sec);
- queue_delayed_work(dev->sriov.alias_guid.ports_guid[port_index].wq,
- &dev->sriov.alias_guid.ports_guid[port_index].
+ queue_delayed_work(dev->sriov->alias_guid.ports_guid[port_index].wq,
+ &dev->sriov->alias_guid.ports_guid[port_index].
alias_guid_work,
msecs_to_jiffies(resched_delay_sec * 1000));
}
@@ -445,8 +445,8 @@ static void aliasguid_query_handler(int status,
kfree(cb_ctx);
} else
complete(&cb_ctx->done);
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
}
static void invalidate_guid_record(struct mlx4_ib_dev *dev, u8 port, int index)
@@ -455,13 +455,13 @@ static void invalidate_guid_record(struct mlx4_ib_dev *dev, u8 port, int index)
u64 cur_admin_val;
ib_sa_comp_mask comp_mask = 0;
- dev->sriov.alias_guid.ports_guid[port - 1].all_rec_per_port[index].status
+ dev->sriov->alias_guid.ports_guid[port - 1].all_rec_per_port[index].status
= MLX4_GUID_INFO_STATUS_SET;
/* calculate the comp_mask for that record.*/
for (i = 0; i < NUM_ALIAS_GUID_IN_REC; i++) {
cur_admin_val =
- *(u64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
+ *(u64 *)&dev->sriov->alias_guid.ports_guid[port - 1].
all_rec_per_port[index].all_recs[GUID_REC_SIZE * i];
/*
check the admin value: if it's for delete (~00LL) or
@@ -474,11 +474,11 @@ static void invalidate_guid_record(struct mlx4_ib_dev *dev, u8 port, int index)
continue;
comp_mask |= mlx4_ib_get_aguid_comp_mask_from_ix(i);
}
- dev->sriov.alias_guid.ports_guid[port - 1].
+ dev->sriov->alias_guid.ports_guid[port - 1].
all_rec_per_port[index].guid_indexes |= comp_mask;
- if (dev->sriov.alias_guid.ports_guid[port - 1].
+ if (dev->sriov->alias_guid.ports_guid[port - 1].
all_rec_per_port[index].guid_indexes)
- dev->sriov.alias_guid.ports_guid[port - 1].
+ dev->sriov->alias_guid.ports_guid[port - 1].
all_rec_per_port[index].status = MLX4_GUID_INFO_STATUS_IDLE;
}
@@ -497,7 +497,7 @@ static int set_guid_rec(struct ib_device *ibdev,
int index = rec->block_num;
struct mlx4_sriov_alias_guid_info_rec_det *rec_det = &rec->rec_det;
struct list_head *head =
- &dev->sriov.alias_guid.ports_guid[port - 1].cb_list;
+ &dev->sriov->alias_guid.ports_guid[port - 1].cb_list;
memset(&attr, 0, sizeof(attr));
err = __mlx4_ib_query_port(ibdev, port, &attr, 1);
@@ -537,12 +537,12 @@ static int set_guid_rec(struct ib_device *ibdev,
rec_det->guid_indexes;
init_completion(&callback_context->done);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
list_add_tail(&callback_context->list, head);
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
callback_context->query_id =
- ib_sa_guid_info_rec_query(dev->sriov.alias_guid.sa_client,
+ ib_sa_guid_info_rec_query(dev->sriov->alias_guid.sa_client,
ibdev, port, &guid_info_rec,
comp_mask, rec->method, 1000,
GFP_KERNEL, aliasguid_query_handler,
@@ -552,10 +552,10 @@ static int set_guid_rec(struct ib_device *ibdev,
pr_debug("ib_sa_guid_info_rec_query failed, query_id: "
"%d. will reschedule to the next 1 sec.\n",
callback_context->query_id);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
list_del(&callback_context->list);
kfree(callback_context);
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
resched_delay = 1 * HZ;
err = -EAGAIN;
goto new_schedule;
@@ -564,16 +564,16 @@ static int set_guid_rec(struct ib_device *ibdev,
goto out;
new_schedule:
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
invalidate_guid_record(dev, port, index);
- if (!dev->sriov.is_going_down) {
- queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
- &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
+ if (!dev->sriov->is_going_down) {
+ queue_delayed_work(dev->sriov->alias_guid.ports_guid[port - 1].wq,
+ &dev->sriov->alias_guid.ports_guid[port - 1].alias_guid_work,
resched_delay);
}
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
out:
return err;
@@ -593,7 +593,7 @@ static void mlx4_ib_guid_port_init(struct mlx4_ib_dev *dev, int port)
!mlx4_is_slave_active(dev->dev, entry))
continue;
guid = mlx4_get_admin_guid(dev->dev, entry, port);
- *(__be64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
+ *(__be64 *)&dev->sriov->alias_guid.ports_guid[port - 1].
all_rec_per_port[j].all_recs
[GUID_REC_SIZE * k] = guid;
pr_debug("guid was set, entry=%d, val=0x%llx, port=%d\n",
@@ -610,32 +610,32 @@ void mlx4_ib_invalidate_all_guid_record(struct mlx4_ib_dev *dev, int port)
pr_debug("port %d\n", port);
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
- if (dev->sriov.alias_guid.ports_guid[port - 1].state_flags &
+ if (dev->sriov->alias_guid.ports_guid[port - 1].state_flags &
GUID_STATE_NEED_PORT_INIT) {
mlx4_ib_guid_port_init(dev, port);
- dev->sriov.alias_guid.ports_guid[port - 1].state_flags &=
+ dev->sriov->alias_guid.ports_guid[port - 1].state_flags &=
(~GUID_STATE_NEED_PORT_INIT);
}
for (i = 0; i < NUM_ALIAS_GUID_REC_IN_PORT; i++)
invalidate_guid_record(dev, port, i);
- if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) {
+ if (mlx4_is_master(dev->dev) && !dev->sriov->is_going_down) {
/*
make sure no work waits in the queue, if the work is already
queued(not on the timer) the cancel will fail. That is not a problem
because we just want the work started.
*/
- cancel_delayed_work(&dev->sriov.alias_guid.
+ cancel_delayed_work(&dev->sriov->alias_guid.
ports_guid[port - 1].alias_guid_work);
- queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
- &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
+ queue_delayed_work(dev->sriov->alias_guid.ports_guid[port - 1].wq,
+ &dev->sriov->alias_guid.ports_guid[port - 1].alias_guid_work,
0);
}
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
}
static void set_required_record(struct mlx4_ib_dev *dev, u8 port,
@@ -648,7 +648,7 @@ static void set_required_record(struct mlx4_ib_dev *dev, u8 port,
ib_sa_comp_mask delete_guid_indexes = 0;
ib_sa_comp_mask set_guid_indexes = 0;
struct mlx4_sriov_alias_guid_info_rec_det *rec =
- &dev->sriov.alias_guid.ports_guid[port].
+ &dev->sriov->alias_guid.ports_guid[port].
all_rec_per_port[record_index];
for (i = 0; i < NUM_ALIAS_GUID_IN_REC; i++) {
@@ -697,7 +697,7 @@ static int get_low_record_time_index(struct mlx4_ib_dev *dev, u8 port,
int j;
for (j = 0; j < NUM_ALIAS_GUID_REC_IN_PORT; j++) {
- rec = dev->sriov.alias_guid.ports_guid[port].
+ rec = dev->sriov->alias_guid.ports_guid[port].
all_rec_per_port[j];
if (rec.status == MLX4_GUID_INFO_STATUS_IDLE &&
rec.guid_indexes) {
@@ -727,7 +727,7 @@ static int get_next_record_to_update(struct mlx4_ib_dev *dev, u8 port,
int record_index;
int ret = 0;
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags);
record_index = get_low_record_time_index(dev, port, NULL);
if (record_index < 0) {
@@ -737,7 +737,7 @@ static int get_next_record_to_update(struct mlx4_ib_dev *dev, u8 port,
set_required_record(dev, port, rec, record_index);
out:
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags);
return ret;
}
@@ -753,7 +753,7 @@ static void alias_guid_work(struct work_struct *work)
struct mlx4_ib_sriov *ib_sriov = container_of(sriov_alias_guid,
struct mlx4_ib_sriov,
alias_guid);
- struct mlx4_ib_dev *dev = container_of(ib_sriov, struct mlx4_ib_dev, sriov);
+ struct mlx4_ib_dev *dev = ib_sriov->parent;
rec = kzalloc(sizeof *rec, GFP_KERNEL);
if (!rec)
@@ -778,33 +778,33 @@ void mlx4_ib_init_alias_guid_work(struct mlx4_ib_dev *dev, int port)
if (!mlx4_is_master(dev->dev))
return;
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
- if (!dev->sriov.is_going_down) {
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ spin_lock_irqsave(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ if (!dev->sriov->is_going_down) {
/* If there is pending one should cancel then run, otherwise
* won't run till previous one is ended as same work
* struct is used.
*/
- cancel_delayed_work(&dev->sriov.alias_guid.ports_guid[port].
+ cancel_delayed_work(&dev->sriov->alias_guid.ports_guid[port].
alias_guid_work);
- queue_delayed_work(dev->sriov.alias_guid.ports_guid[port].wq,
- &dev->sriov.alias_guid.ports_guid[port].alias_guid_work, 0);
+ queue_delayed_work(dev->sriov->alias_guid.ports_guid[port].wq,
+ &dev->sriov->alias_guid.ports_guid[port].alias_guid_work, 0);
}
- spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->alias_guid.ag_work_lock, flags1);
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
}
void mlx4_ib_destroy_alias_guid_service(struct mlx4_ib_dev *dev)
{
int i;
- struct mlx4_ib_sriov *sriov = &dev->sriov;
+ struct mlx4_ib_sriov *sriov = dev->sriov;
struct mlx4_alias_guid_work_context *cb_ctx;
struct mlx4_sriov_alias_guid_port_rec_det *det;
struct ib_sa_query *sa_query;
unsigned long flags;
for (i = 0 ; i < dev->num_ports; i++) {
- cancel_delayed_work(&dev->sriov.alias_guid.ports_guid[i].alias_guid_work);
+ cancel_delayed_work(&dev->sriov->alias_guid.ports_guid[i].alias_guid_work);
det = &sriov->alias_guid.ports_guid[i];
spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags);
while (!list_empty(&det->cb_list)) {
@@ -823,11 +823,11 @@ void mlx4_ib_destroy_alias_guid_service(struct mlx4_ib_dev *dev)
spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags);
}
for (i = 0 ; i < dev->num_ports; i++) {
- flush_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
- destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
+ flush_workqueue(dev->sriov->alias_guid.ports_guid[i].wq);
+ destroy_workqueue(dev->sriov->alias_guid.ports_guid[i].wq);
}
- ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
- kfree(dev->sriov.alias_guid.sa_client);
+ ib_sa_unregister_client(dev->sriov->alias_guid.sa_client);
+ kfree(dev->sriov->alias_guid.sa_client);
}
int mlx4_ib_init_alias_guid_service(struct mlx4_ib_dev *dev)
@@ -839,14 +839,14 @@ int mlx4_ib_init_alias_guid_service(struct mlx4_ib_dev *dev)
if (!mlx4_is_master(dev->dev))
return 0;
- dev->sriov.alias_guid.sa_client =
- kzalloc(sizeof *dev->sriov.alias_guid.sa_client, GFP_KERNEL);
- if (!dev->sriov.alias_guid.sa_client)
+ dev->sriov->alias_guid.sa_client =
+ kzalloc(sizeof *dev->sriov->alias_guid.sa_client, GFP_KERNEL);
+ if (!dev->sriov->alias_guid.sa_client)
return -ENOMEM;
- ib_sa_register_client(dev->sriov.alias_guid.sa_client);
+ ib_sa_register_client(dev->sriov->alias_guid.sa_client);
- spin_lock_init(&dev->sriov.alias_guid.ag_work_lock);
+ spin_lock_init(&dev->sriov->alias_guid.ag_work_lock);
for (i = 1; i <= dev->num_ports; ++i) {
if (dev->ib_dev.query_gid(&dev->ib_dev , i, 0, &gid)) {
@@ -856,18 +856,18 @@ int mlx4_ib_init_alias_guid_service(struct mlx4_ib_dev *dev)
}
for (i = 0 ; i < dev->num_ports; i++) {
- memset(&dev->sriov.alias_guid.ports_guid[i], 0,
+ memset(&dev->sriov->alias_guid.ports_guid[i], 0,
sizeof (struct mlx4_sriov_alias_guid_port_rec_det));
- dev->sriov.alias_guid.ports_guid[i].state_flags |=
+ dev->sriov->alias_guid.ports_guid[i].state_flags |=
GUID_STATE_NEED_PORT_INIT;
for (j = 0; j < NUM_ALIAS_GUID_REC_IN_PORT; j++) {
/* mark each val as it was deleted */
- memset(dev->sriov.alias_guid.ports_guid[i].
+ memset(dev->sriov->alias_guid.ports_guid[i].
all_rec_per_port[j].all_recs, 0xFF,
- sizeof(dev->sriov.alias_guid.ports_guid[i].
+ sizeof(dev->sriov->alias_guid.ports_guid[i].
all_rec_per_port[j].all_recs));
}
- INIT_LIST_HEAD(&dev->sriov.alias_guid.ports_guid[i].cb_list);
+ INIT_LIST_HEAD(&dev->sriov->alias_guid.ports_guid[i].cb_list);
/*prepare the records, set them to be allocated by sm*/
if (mlx4_ib_sm_guid_assign)
for (j = 1; j < NUM_ALIAS_GUID_PER_PORT; j++)
@@ -875,31 +875,31 @@ int mlx4_ib_init_alias_guid_service(struct mlx4_ib_dev *dev)
for (j = 0 ; j < NUM_ALIAS_GUID_REC_IN_PORT; j++)
invalidate_guid_record(dev, i + 1, j);
- dev->sriov.alias_guid.ports_guid[i].parent = &dev->sriov.alias_guid;
- dev->sriov.alias_guid.ports_guid[i].port = i;
+ dev->sriov->alias_guid.ports_guid[i].parent = &dev->sriov->alias_guid;
+ dev->sriov->alias_guid.ports_guid[i].port = i;
snprintf(alias_wq_name, sizeof alias_wq_name, "alias_guid%d", i);
- dev->sriov.alias_guid.ports_guid[i].wq =
+ dev->sriov->alias_guid.ports_guid[i].wq =
alloc_ordered_workqueue(alias_wq_name, WQ_MEM_RECLAIM);
- if (!dev->sriov.alias_guid.ports_guid[i].wq) {
+ if (!dev->sriov->alias_guid.ports_guid[i].wq) {
ret = -ENOMEM;
goto err_thread;
}
- INIT_DELAYED_WORK(&dev->sriov.alias_guid.ports_guid[i].alias_guid_work,
+ INIT_DELAYED_WORK(&dev->sriov->alias_guid.ports_guid[i].alias_guid_work,
alias_guid_work);
}
return 0;
err_thread:
for (--i; i >= 0; i--) {
- destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
- dev->sriov.alias_guid.ports_guid[i].wq = NULL;
+ destroy_workqueue(dev->sriov->alias_guid.ports_guid[i].wq);
+ dev->sriov->alias_guid.ports_guid[i].wq = NULL;
}
err_unregister:
- ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
- kfree(dev->sriov.alias_guid.sa_client);
- dev->sriov.alias_guid.sa_client = NULL;
+ ib_sa_unregister_client(dev->sriov->alias_guid.sa_client);
+ kfree(dev->sriov->alias_guid.sa_client);
+ dev->sriov->alias_guid.sa_client = NULL;
pr_err("init_alias_guid_service: Failed. (ret:%d)\n", ret);
return ret;
}
diff --git a/drivers/infiniband/hw/mlx4/cm.c b/drivers/infiniband/hw/mlx4/cm.c
index fedaf82..3712109 100644
--- a/drivers/infiniband/hw/mlx4/cm.c
+++ b/drivers/infiniband/hw/mlx4/cm.c
@@ -143,7 +143,7 @@ static union ib_gid gid_from_req_msg(struct ib_device *ibdev, struct ib_mad *mad
static struct id_map_entry *
id_map_find_by_sl_id(struct ib_device *ibdev, u32 slave_id, u32 sl_cm_id)
{
- struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map;
+ struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov->sl_id_map;
struct rb_node *node = sl_id_map->rb_node;
while (node) {
@@ -170,7 +170,7 @@ static void id_map_ent_timeout(struct work_struct *work)
struct id_map_entry *ent = container_of(delay, struct id_map_entry, timeout);
struct id_map_entry *db_ent, *found_ent;
struct mlx4_ib_dev *dev = ent->dev;
- struct mlx4_ib_sriov *sriov = &dev->sriov;
+ struct mlx4_ib_sriov *sriov = dev->sriov;
struct rb_root *sl_id_map = &sriov->sl_id_map;
int pv_id = (int) ent->pv_cm_id;
@@ -191,7 +191,7 @@ static void id_map_ent_timeout(struct work_struct *work)
static void id_map_find_del(struct ib_device *ibdev, int pv_cm_id)
{
- struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
+ struct mlx4_ib_sriov *sriov = to_mdev(ibdev)->sriov;
struct rb_root *sl_id_map = &sriov->sl_id_map;
struct id_map_entry *ent, *found_ent;
@@ -209,7 +209,7 @@ static void id_map_find_del(struct ib_device *ibdev, int pv_cm_id)
static void sl_id_map_add(struct ib_device *ibdev, struct id_map_entry *new)
{
- struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map;
+ struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov->sl_id_map;
struct rb_node **link = &sl_id_map->rb_node, *parent = NULL;
struct id_map_entry *ent;
int slave_id = new->slave_id;
@@ -244,7 +244,7 @@ id_map_alloc(struct ib_device *ibdev, int slave_id, u32 sl_cm_id)
{
int ret;
struct id_map_entry *ent;
- struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
+ struct mlx4_ib_sriov *sriov = to_mdev(ibdev)->sriov;
ent = kmalloc(sizeof (struct id_map_entry), GFP_KERNEL);
if (!ent)
@@ -257,7 +257,7 @@ id_map_alloc(struct ib_device *ibdev, int slave_id, u32 sl_cm_id)
INIT_DELAYED_WORK(&ent->timeout, id_map_ent_timeout);
idr_preload(GFP_KERNEL);
- spin_lock(&to_mdev(ibdev)->sriov.id_map_lock);
+ spin_lock(&to_mdev(ibdev)->sriov->id_map_lock);
ret = idr_alloc_cyclic(&sriov->pv_id_table, ent, 0, 0, GFP_NOWAIT);
if (ret >= 0) {
@@ -282,7 +282,7 @@ static struct id_map_entry *
id_map_get(struct ib_device *ibdev, int *pv_cm_id, int slave_id, int sl_cm_id)
{
struct id_map_entry *ent;
- struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
+ struct mlx4_ib_sriov *sriov = to_mdev(ibdev)->sriov;
spin_lock(&sriov->id_map_lock);
if (*pv_cm_id == -1) {
@@ -298,7 +298,7 @@ id_map_get(struct ib_device *ibdev, int *pv_cm_id, int slave_id, int sl_cm_id)
static void schedule_delayed(struct ib_device *ibdev, struct id_map_entry *id)
{
- struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
+ struct mlx4_ib_sriov *sriov = to_mdev(ibdev)->sriov;
unsigned long flags;
spin_lock(&sriov->id_map_lock);
@@ -404,17 +404,17 @@ int mlx4_ib_demux_cm_handler(struct ib_device *ibdev, int port, int *slave,
void mlx4_ib_cm_paravirt_init(struct mlx4_ib_dev *dev)
{
- spin_lock_init(&dev->sriov.id_map_lock);
- INIT_LIST_HEAD(&dev->sriov.cm_list);
- dev->sriov.sl_id_map = RB_ROOT;
- idr_init(&dev->sriov.pv_id_table);
+ spin_lock_init(&dev->sriov->id_map_lock);
+ INIT_LIST_HEAD(&dev->sriov->cm_list);
+ dev->sriov->sl_id_map = RB_ROOT;
+ idr_init(&dev->sriov->pv_id_table);
}
/* slave = -1 ==> all slaves */
/* TBD -- call paravirt clean for single slave. Need for slave RESET event */
void mlx4_ib_cm_paravirt_clean(struct mlx4_ib_dev *dev, int slave)
{
- struct mlx4_ib_sriov *sriov = &dev->sriov;
+ struct mlx4_ib_sriov *sriov = dev->sriov;
struct rb_root *sl_id_map = &sriov->sl_id_map;
struct list_head lh;
struct rb_node *nd;
@@ -423,7 +423,7 @@ void mlx4_ib_cm_paravirt_clean(struct mlx4_ib_dev *dev, int slave)
/* cancel all delayed work queue entries */
INIT_LIST_HEAD(&lh);
spin_lock(&sriov->id_map_lock);
- list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) {
+ list_for_each_entry_safe(map, tmp_map, &dev->sriov->cm_list, list) {
if (slave < 0 || slave == map->slave_id) {
if (map->scheduled_delete)
need_flush |= !cancel_delayed_work(&map->timeout);
@@ -446,7 +446,7 @@ void mlx4_ib_cm_paravirt_clean(struct mlx4_ib_dev *dev, int slave)
rb_erase(&ent->node, sl_id_map);
idr_remove(&sriov->pv_id_table, (int) ent->pv_cm_id);
}
- list_splice_init(&dev->sriov.cm_list, &lh);
+ list_splice_init(&dev->sriov->cm_list, &lh);
} else {
/* first, move nodes belonging to slave to db remove list */
nd = rb_first(sl_id_map);
@@ -464,7 +464,7 @@ void mlx4_ib_cm_paravirt_clean(struct mlx4_ib_dev *dev, int slave)
}
/* add remaining nodes from cm_list */
- list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) {
+ list_for_each_entry_safe(map, tmp_map, &dev->sriov->cm_list, list) {
if (slave == map->slave_id)
list_move_tail(&map->list, &lh);
}
diff --git a/drivers/infiniband/hw/mlx4/mad.c b/drivers/infiniband/hw/mlx4/mad.c
index 3eceb46..88fe6cd 100644
--- a/drivers/infiniband/hw/mlx4/mad.c
+++ b/drivers/infiniband/hw/mlx4/mad.c
@@ -281,7 +281,7 @@ static void smp_snoop(struct ib_device *ibdev, u8 port_num, const struct ib_mad
if (pkey_change_bitmap) {
mlx4_ib_dispatch_event(dev, port_num,
IB_EVENT_PKEY_CHANGE);
- if (!dev->sriov.is_going_down)
+ if (!dev->sriov->is_going_down)
__propagate_pkey_ev(dev, port_num, bn,
pkey_change_bitmap);
}
@@ -296,7 +296,7 @@ static void smp_snoop(struct ib_device *ibdev, u8 port_num, const struct ib_mad
IB_EVENT_GID_CHANGE);
/*if master, notify relevant slaves*/
if (mlx4_is_master(dev->dev) &&
- !dev->sriov.is_going_down) {
+ !dev->sriov->is_going_down) {
bn = be32_to_cpu(((struct ib_smp *)mad)->attr_mod);
mlx4_ib_update_cache_on_guid_change(dev, bn, port_num,
(u8 *)(&((struct ib_smp *)mad)->data));
@@ -435,7 +435,7 @@ int mlx4_ib_find_real_gid(struct ib_device *ibdev, u8 port, __be64 guid)
int i;
for (i = 0; i < dev->dev->caps.sqp_demux; i++) {
- if (dev->sriov.demux[port - 1].guid_cache[i] == guid)
+ if (dev->sriov->demux[port - 1].guid_cache[i] == guid)
return i;
}
return -1;
@@ -523,7 +523,7 @@ int mlx4_ib_send_to_slave(struct mlx4_ib_dev *dev, int slave, u8 port,
if (dest_qpt > IB_QPT_GSI)
return -EINVAL;
- tun_ctx = dev->sriov.demux[port-1].tun[slave];
+ tun_ctx = dev->sriov->demux[port-1].tun[slave];
/* check if proxy qp created */
if (!tun_ctx || tun_ctx->state != DEMUX_PV_STATE_ACTIVE)
@@ -736,7 +736,7 @@ static int mlx4_ib_demux_mad(struct ib_device *ibdev, u8 port,
if (grh->dgid.global.interface_id ==
cpu_to_be64(IB_SA_WELL_KNOWN_GUID) &&
grh->dgid.global.subnet_prefix == cpu_to_be64(
- atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix))) {
+ atomic64_read(&dev->sriov->demux[port - 1].subnet_prefix))) {
slave = 0;
} else {
slave = mlx4_ib_find_real_gid(ibdev, port,
@@ -1085,7 +1085,7 @@ static void handle_lid_change_event(struct mlx4_ib_dev *dev, u8 port_num)
{
mlx4_ib_dispatch_event(dev, port_num, IB_EVENT_LID_CHANGE);
- if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down)
+ if (mlx4_is_master(dev->dev) && !dev->sriov->is_going_down)
mlx4_gen_slaves_port_mgt_ev(dev->dev, port_num,
MLX4_EQ_PORT_INFO_LID_CHANGE_MASK);
}
@@ -1096,8 +1096,8 @@ static void handle_client_rereg_event(struct mlx4_ib_dev *dev, u8 port_num)
if (mlx4_is_master(dev->dev)) {
mlx4_ib_invalidate_all_guid_record(dev, port_num);
- if (!dev->sriov.is_going_down) {
- mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0);
+ if (!dev->sriov->is_going_down) {
+ mlx4_ib_mcg_port_cleanup(&dev->sriov->demux[port_num - 1], 0);
mlx4_gen_slaves_port_mgt_ev(dev->dev, port_num,
MLX4_EQ_PORT_INFO_CLIENT_REREG_MASK);
}
@@ -1223,9 +1223,9 @@ void handle_port_mgmt_change_event(struct work_struct *work)
} else {
pr_debug("Changing QP1 subnet prefix for port %d. old=0x%llx. new=0x%llx\n",
port,
- (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix),
+ (u64)atomic64_read(&dev->sriov->demux[port - 1].subnet_prefix),
be64_to_cpu(gid.global.subnet_prefix));
- atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix,
+ atomic64_set(&dev->sriov->demux[port - 1].subnet_prefix,
be64_to_cpu(gid.global.subnet_prefix));
}
}
@@ -1242,7 +1242,7 @@ void handle_port_mgmt_change_event(struct work_struct *work)
case MLX4_DEV_PMC_SUBTYPE_PKEY_TABLE:
mlx4_ib_dispatch_event(dev, port, IB_EVENT_PKEY_CHANGE);
- if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down)
+ if (mlx4_is_master(dev->dev) && !dev->sriov->is_going_down)
propagate_pkey_ev(dev, port, eqe);
break;
case MLX4_DEV_PMC_SUBTYPE_GUID_INFO:
@@ -1250,7 +1250,7 @@ void handle_port_mgmt_change_event(struct work_struct *work)
if (!mlx4_is_master(dev->dev))
mlx4_ib_dispatch_event(dev, port, IB_EVENT_GID_CHANGE);
/*if master, notify relevant slaves*/
- else if (!dev->sriov.is_going_down) {
+ else if (!dev->sriov->is_going_down) {
tbl_block = GET_BLK_PTR_FROM_EQE(eqe);
change_bitmap = GET_MASK_FROM_EQE(eqe);
handle_slaves_guid_change(dev, port, tbl_block, change_bitmap);
@@ -1299,10 +1299,10 @@ static void mlx4_ib_tunnel_comp_handler(struct ib_cq *cq, void *arg)
unsigned long flags;
struct mlx4_ib_demux_pv_ctx *ctx = cq->cq_context;
struct mlx4_ib_dev *dev = to_mdev(ctx->ib_dev);
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- if (!dev->sriov.is_going_down && ctx->state == DEMUX_PV_STATE_ACTIVE)
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ if (!dev->sriov->is_going_down && ctx->state == DEMUX_PV_STATE_ACTIVE)
queue_work(ctx->wq, &ctx->work);
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
}
static int mlx4_ib_post_pv_qp_buf(struct mlx4_ib_demux_pv_ctx *ctx,
@@ -1373,7 +1373,7 @@ int mlx4_ib_send_to_wire(struct mlx4_ib_dev *dev, int slave, u8 port,
u16 wire_pkey_ix;
int src_qpnum;
- sqp_ctx = dev->sriov.sqps[port-1];
+ sqp_ctx = dev->sriov->sqps[port-1];
/* check if proxy qp created */
if (!sqp_ctx || sqp_ctx->state != DEMUX_PV_STATE_ACTIVE)
@@ -1960,9 +1960,9 @@ static int alloc_pv_object(struct mlx4_ib_dev *dev, int slave, int port,
static void free_pv_object(struct mlx4_ib_dev *dev, int slave, int port)
{
- if (dev->sriov.demux[port - 1].tun[slave]) {
- kfree(dev->sriov.demux[port - 1].tun[slave]);
- dev->sriov.demux[port - 1].tun[slave] = NULL;
+ if (dev->sriov->demux[port - 1].tun[slave]) {
+ kfree(dev->sriov->demux[port - 1].tun[slave]);
+ dev->sriov->demux[port - 1].tun[slave] = NULL;
}
}
@@ -2036,7 +2036,7 @@ static int create_pv_resources(struct ib_device *ibdev, int slave, int port,
else
INIT_WORK(&ctx->work, mlx4_ib_sqp_comp_worker);
- ctx->wq = to_mdev(ibdev)->sriov.demux[port - 1].wq;
+ ctx->wq = to_mdev(ibdev)->sriov->demux[port - 1].wq;
ret = ib_req_notify_cq(ctx->cq, IB_CQ_NEXT_COMP);
if (ret) {
@@ -2107,25 +2107,25 @@ static int mlx4_ib_tunnels_update(struct mlx4_ib_dev *dev, int slave,
int ret = 0;
if (!do_init) {
- clean_vf_mcast(&dev->sriov.demux[port - 1], slave);
+ clean_vf_mcast(&dev->sriov->demux[port - 1], slave);
/* for master, destroy real sqp resources */
if (slave == mlx4_master_func_num(dev->dev))
destroy_pv_resources(dev, slave, port,
- dev->sriov.sqps[port - 1], 1);
+ dev->sriov->sqps[port - 1], 1);
/* destroy the tunnel qp resources */
destroy_pv_resources(dev, slave, port,
- dev->sriov.demux[port - 1].tun[slave], 1);
+ dev->sriov->demux[port - 1].tun[slave], 1);
return 0;
}
/* create the tunnel qp resources */
ret = create_pv_resources(&dev->ib_dev, slave, port, 1,
- dev->sriov.demux[port - 1].tun[slave]);
+ dev->sriov->demux[port - 1].tun[slave]);
/* for master, create the real sqp resources */
if (!ret && slave == mlx4_master_func_num(dev->dev))
ret = create_pv_resources(&dev->ib_dev, slave, port, 0,
- dev->sriov.sqps[port - 1]);
+ dev->sriov->sqps[port - 1]);
return ret;
}
@@ -2276,8 +2276,8 @@ int mlx4_ib_init_sriov(struct mlx4_ib_dev *dev)
if (!mlx4_is_mfunc(dev->dev))
return 0;
- dev->sriov.is_going_down = 0;
- spin_lock_init(&dev->sriov.going_down_lock);
+ dev->sriov->is_going_down = 0;
+ spin_lock_init(&dev->sriov->going_down_lock);
mlx4_ib_cm_paravirt_init(dev);
mlx4_ib_warn(&dev->ib_dev, "multi-function enabled\n");
@@ -2312,14 +2312,14 @@ int mlx4_ib_init_sriov(struct mlx4_ib_dev *dev)
err = __mlx4_ib_query_gid(&dev->ib_dev, i + 1, 0, &gid, 1);
if (err)
goto demux_err;
- dev->sriov.demux[i].guid_cache[0] = gid.global.interface_id;
- atomic64_set(&dev->sriov.demux[i].subnet_prefix,
+ dev->sriov->demux[i].guid_cache[0] = gid.global.interface_id;
+ atomic64_set(&dev->sriov->demux[i].subnet_prefix,
be64_to_cpu(gid.global.subnet_prefix));
err = alloc_pv_object(dev, mlx4_master_func_num(dev->dev), i + 1,
- &dev->sriov.sqps[i]);
+ &dev->sriov->sqps[i]);
if (err)
goto demux_err;
- err = mlx4_ib_alloc_demux_ctx(dev, &dev->sriov.demux[i], i + 1);
+ err = mlx4_ib_alloc_demux_ctx(dev, &dev->sriov->demux[i], i + 1);
if (err)
goto free_pv;
}
@@ -2331,7 +2331,7 @@ int mlx4_ib_init_sriov(struct mlx4_ib_dev *dev)
demux_err:
while (--i >= 0) {
free_pv_object(dev, mlx4_master_func_num(dev->dev), i + 1);
- mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]);
+ mlx4_ib_free_demux_ctx(&dev->sriov->demux[i]);
}
mlx4_ib_device_unregister_sysfs(dev);
@@ -2352,16 +2352,16 @@ void mlx4_ib_close_sriov(struct mlx4_ib_dev *dev)
if (!mlx4_is_mfunc(dev->dev))
return;
- spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
- dev->sriov.is_going_down = 1;
- spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
+ spin_lock_irqsave(&dev->sriov->going_down_lock, flags);
+ dev->sriov->is_going_down = 1;
+ spin_unlock_irqrestore(&dev->sriov->going_down_lock, flags);
if (mlx4_is_master(dev->dev)) {
for (i = 0; i < dev->num_ports; i++) {
- flush_workqueue(dev->sriov.demux[i].ud_wq);
- mlx4_ib_free_sqp_ctx(dev->sriov.sqps[i]);
- kfree(dev->sriov.sqps[i]);
- dev->sriov.sqps[i] = NULL;
- mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]);
+ flush_workqueue(dev->sriov->demux[i].ud_wq);
+ mlx4_ib_free_sqp_ctx(dev->sriov->sqps[i]);
+ kfree(dev->sriov->sqps[i]);
+ dev->sriov->sqps[i] = NULL;
+ mlx4_ib_free_demux_ctx(&dev->sriov->demux[i]);
}
mlx4_ib_cm_paravirt_clean(dev, -1);
diff --git a/drivers/infiniband/hw/mlx4/main.c b/drivers/infiniband/hw/mlx4/main.c
index 8ba0103..6d1483d 100644
--- a/drivers/infiniband/hw/mlx4/main.c
+++ b/drivers/infiniband/hw/mlx4/main.c
@@ -2596,6 +2596,7 @@ static void mlx4_ib_release(struct ib_device *device)
kvfree(ibdev->iboe);
kvfree(ibdev->pkeys);
+ kvfree(ibdev->sriov);
}
static void *mlx4_ib_add(struct mlx4_dev *dev)
@@ -2641,6 +2642,12 @@ static void *mlx4_ib_add(struct mlx4_dev *dev)
if (!ibdev->pkeys)
goto err_dealloc;
+ ibdev->sriov = kvzalloc(sizeof(struct mlx4_ib_sriov), GFP_KERNEL);
+ if (!ibdev->sriov)
+ goto err_dealloc;
+
+ ibdev->sriov->parent = ibdev;
+
if (mlx4_pd_alloc(dev, &ibdev->priv_pdn))
goto err_dealloc;
@@ -3152,13 +3159,13 @@ static void do_slave_init(struct mlx4_ib_dev *ibdev, int slave, int do_init)
dm[i]->dev = ibdev;
}
/* initialize or tear down tunnel QPs for the slave */
- spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags);
- if (!ibdev->sriov.is_going_down) {
+ spin_lock_irqsave(&ibdev->sriov->going_down_lock, flags);
+ if (!ibdev->sriov->is_going_down) {
for (i = 0; i < ports; i++)
- queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work);
- spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags);
+ queue_work(ibdev->sriov->demux[i].ud_wq, &dm[i]->work);
+ spin_unlock_irqrestore(&ibdev->sriov->going_down_lock, flags);
} else {
- spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags);
+ spin_unlock_irqrestore(&ibdev->sriov->going_down_lock, flags);
for (i = 0; i < ports; i++)
kfree(dm[i]);
}
diff --git a/drivers/infiniband/hw/mlx4/mcg.c b/drivers/infiniband/hw/mlx4/mcg.c
index 81ffc00..6415326 100644
--- a/drivers/infiniband/hw/mlx4/mcg.c
+++ b/drivers/infiniband/hw/mlx4/mcg.c
@@ -884,7 +884,7 @@ int mlx4_ib_mcg_demux_handler(struct ib_device *ibdev, int port, int slave,
{
struct mlx4_ib_dev *dev = to_mdev(ibdev);
struct ib_sa_mcmember_data *rec = (struct ib_sa_mcmember_data *)mad->data;
- struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1];
+ struct mlx4_ib_demux_ctx *ctx = &dev->sriov->demux[port - 1];
struct mcast_group *group;
switch (mad->mad_hdr.method) {
@@ -933,7 +933,7 @@ int mlx4_ib_mcg_multiplex_handler(struct ib_device *ibdev, int port,
{
struct mlx4_ib_dev *dev = to_mdev(ibdev);
struct ib_sa_mcmember_data *rec = (struct ib_sa_mcmember_data *)sa_mad->data;
- struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1];
+ struct mlx4_ib_demux_ctx *ctx = &dev->sriov->demux[port - 1];
struct mcast_group *group;
struct mcast_req *req;
int may_create = 0;
diff --git a/drivers/infiniband/hw/mlx4/mlx4_ib.h b/drivers/infiniband/hw/mlx4/mlx4_ib.h
index 2b5a9b2..dfe3a5c 100644
--- a/drivers/infiniband/hw/mlx4/mlx4_ib.h
+++ b/drivers/infiniband/hw/mlx4/mlx4_ib.h
@@ -501,6 +501,7 @@ struct mlx4_ib_sriov {
spinlock_t id_map_lock;
struct rb_root sl_id_map;
struct idr pv_id_table;
+ struct mlx4_ib_dev *parent;
};
struct gid_cache_context {
@@ -597,7 +598,7 @@ struct mlx4_ib_dev {
struct ib_ah *sm_ah[MLX4_MAX_PORTS];
spinlock_t sm_lock;
atomic64_t sl2vl[MLX4_MAX_PORTS];
- struct mlx4_ib_sriov sriov;
+ struct mlx4_ib_sriov *sriov;
struct mutex cap_mask_mutex;
bool ib_active;
diff --git a/drivers/infiniband/hw/mlx4/qp.c b/drivers/infiniband/hw/mlx4/qp.c
index 853ef6f..f001e2f 100644
--- a/drivers/infiniband/hw/mlx4/qp.c
+++ b/drivers/infiniband/hw/mlx4/qp.c
@@ -3031,11 +3031,11 @@ static int build_mlx_header(struct mlx4_ib_sqp *sqp, const struct ib_ud_wr *wr,
* we must use our own cache
*/
sqp->ud_header.grh.source_gid.global.subnet_prefix =
- cpu_to_be64(atomic64_read(&(to_mdev(ib_dev)->sriov.
+ cpu_to_be64(atomic64_read(&(to_mdev(ib_dev)->sriov->
demux[sqp->qp.port - 1].
subnet_prefix)));
sqp->ud_header.grh.source_gid.global.interface_id =
- to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1].
+ to_mdev(ib_dev)->sriov->demux[sqp->qp.port - 1].
guid_cache[ah->av.ib.gid_index];
} else {
sqp->ud_header.grh.source_gid =
diff --git a/drivers/infiniband/hw/mlx4/sysfs.c b/drivers/infiniband/hw/mlx4/sysfs.c
index a5b4592a..15f7c38 100644
--- a/drivers/infiniband/hw/mlx4/sysfs.c
+++ b/drivers/infiniband/hw/mlx4/sysfs.c
@@ -84,25 +84,25 @@ static ssize_t store_admin_alias_guid(struct device *dev,
pr_err("GUID 0 block 0 is RO\n");
return count;
}
- spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags);
+ spin_lock_irqsave(&mdev->sriov->alias_guid.ag_work_lock, flags);
sscanf(buf, "%llx", &sysadmin_ag_val);
- *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1].
+ *(__be64 *)&mdev->sriov->alias_guid.ports_guid[port->num - 1].
all_rec_per_port[record_num].
all_recs[GUID_REC_SIZE * guid_index_in_rec] =
cpu_to_be64(sysadmin_ag_val);
/* Change the state to be pending for update */
- mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status
+ mdev->sriov->alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status
= MLX4_GUID_INFO_STATUS_IDLE ;
mlx4_set_admin_guid(mdev->dev, cpu_to_be64(sysadmin_ag_val),
mlx4_ib_iov_dentry->entry_num,
port->num);
/* set the record index */
- mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes
+ mdev->sriov->alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes
|= mlx4_ib_get_aguid_comp_mask_from_ix(guid_index_in_rec);
- spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags);
+ spin_unlock_irqrestore(&mdev->sriov->alias_guid.ag_work_lock, flags);
mlx4_ib_init_alias_guid_work(mdev, port->num - 1);
return count;
--
2.1.4
Powered by blists - more mailing lists