[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <47a000b6-8a5f-541f-809d-ad367f81669c@fujitsu.com>
Date: Thu, 17 Nov 2022 13:20:45 +0000
From: "lizhijian@...itsu.com" <lizhijian@...itsu.com>
To: "Daisuke Matsuda (Fujitsu)" <matsuda-daisuke@...itsu.com>,
"linux-rdma@...r.kernel.org" <linux-rdma@...r.kernel.org>,
"leonro@...dia.com" <leonro@...dia.com>,
"jgg@...dia.com" <jgg@...dia.com>,
"zyjzyj2000@...il.com" <zyjzyj2000@...il.com>
CC: "nvdimm@...ts.linux.dev" <nvdimm@...ts.linux.dev>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"rpearsonhpe@...il.com" <rpearsonhpe@...il.com>,
"yangx.jy@...itsu.com" <yangx.jy@...itsu.com>,
"Yasunori Gotou (Fujitsu)" <y-goto@...itsu.com>
Subject: Re: [RFC PATCH v2 1/7] IB/mlx5: Change
ib_umem_odp_map_dma_single_page() to retain umem_mutex
On 11/11/2022 17:22, Daisuke Matsuda wrote:
> ib_umem_odp_map_dma_single_page(), which has been used only by the mlx5
> driver, holds umem_mutex on success and releases on failure. This
> behavior is not convenient for other drivers to use it, so change it to
> always retain mutex on return.
>
> Signed-off-by: Daisuke Matsuda <matsuda-daisuke@...itsu.com>
> ---
> drivers/infiniband/core/umem_odp.c | 8 +++-----
> drivers/infiniband/hw/mlx5/odp.c | 4 +++-
> 2 files changed, 6 insertions(+), 6 deletions(-)
>
> diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c
> index e9fa22d31c23..49da6735f7c8 100644
> --- a/drivers/infiniband/core/umem_odp.c
> +++ b/drivers/infiniband/core/umem_odp.c
> @@ -328,8 +328,8 @@ static int ib_umem_odp_map_dma_single_page(
> *
> * Maps the range passed in the argument to DMA addresses.
> * The DMA addresses of the mapped pages is updated in umem_odp->dma_list.
> - * Upon success the ODP MR will be locked to let caller complete its device
> - * page table update.
> + * The umem mutex is locked in this function. Callers are responsible for
> + * releasing the lock.
> *
> * Returns the number of pages mapped in success, negative error code
> * for failure.
> @@ -453,11 +453,9 @@ int ib_umem_odp_map_dma_and_lock(struct ib_umem_odp *umem_odp, u64 user_virt,
> break;
> }
> }
> - /* upon success lock should stay on hold for the callee */
> +
> if (!ret)
> ret = dma_index - start_idx;
> - else
> - mutex_unlock(&umem_odp->umem_mutex);
>
> out_put_mm:
> mmput_async(owning_mm);
> diff --git a/drivers/infiniband/hw/mlx5/odp.c b/drivers/infiniband/hw/mlx5/odp.c
> index bc97958818bb..a0de27651586 100644
> --- a/drivers/infiniband/hw/mlx5/odp.c
> +++ b/drivers/infiniband/hw/mlx5/odp.c
> @@ -572,8 +572,10 @@ static int pagefault_real_mr(struct mlx5_ib_mr *mr, struct ib_umem_odp *odp,
> access_mask |= ODP_WRITE_ALLOWED_BIT;
>
> np = ib_umem_odp_map_dma_and_lock(odp, user_va, bcnt, access_mask, fault);
> - if (np < 0)
> + if (np < 0) {
> + mutex_unlock(&odp->umem_mutex);
> return np;
> + }
refer to the comments of ib_umem_odp_map_dma_and_lock:
334 * Returns the number of pages mapped in success, negative error
code
335 * for failure.
I don't think it's correct to release the lock in all failure case, for
example when it reaches below error path.
346 int ib_umem_odp_map_dma_and_lock(struct ib_umem_odp *umem_odp, u64
user_virt,
347 u64 bcnt, u64 access_mask, bool
fault)
348 __acquires(&umem_odp->umem_mutex)
349 {
350 struct task_struct *owning_process = NULL;
351 struct mm_struct *owning_mm = umem_odp->umem.owning_mm;
352 int pfn_index, dma_index, ret = 0, start_idx;
353 unsigned int page_shift, hmm_order, pfn_start_idx;
354 unsigned long num_pfns, current_seq;
355 struct hmm_range range = {};
356 unsigned long timeout;
357
358 if (access_mask == 0)
359 return -EINVAL; <<<<< no lock is hold yet
360
361 if (user_virt < ib_umem_start(umem_odp) ||
362 user_virt + bcnt > ib_umem_end(umem_odp))
363 return -EFAULT; <<<<< no lock is hold yet
Further more, you changed public API's the behavior, do it matter for
other out-of-tree drivers which is using it, i'm not familiar with this,
maybe kernel has no restriction on it ?
>
> /*
> * No need to check whether the MTTs really belong to this MR, since
Powered by blists - more mailing lists