lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ca76ef68-5fa4-4a2a-9da2-0378dd7bc2cf@amd.com>
Date:   Tue, 24 Oct 2023 10:45:41 +0200
From:   Christian König <christian.koenig@....com>
To:     Danilo Krummrich <dakr@...hat.com>, airlied@...il.com,
        daniel@...ll.ch, matthew.brost@...el.com,
        thomas.hellstrom@...ux.intel.com, sarah.walker@...tec.com,
        donald.robson@...tec.com, boris.brezillon@...labora.com,
        faith@...strand.net
Cc:     dri-devel@...ts.freedesktop.org, nouveau@...ts.freedesktop.org,
        linux-kernel@...r.kernel.org
Subject: Re: [PATCH drm-misc-next v7 1/7] drm/gpuvm: convert WARN() to
 drm_WARN() variants



Am 23.10.23 um 22:16 schrieb Danilo Krummrich:
> Use drm_WARN() and drm_WARN_ON() variants to indicate drivers the
> context the failing VM resides in.
>
> Signed-off-by: Danilo Krummrich <dakr@...hat.com>
> ---
>   drivers/gpu/drm/drm_gpuvm.c            | 32 ++++++++++++++------------
>   drivers/gpu/drm/nouveau/nouveau_uvmm.c |  3 ++-
>   include/drm/drm_gpuvm.h                |  7 ++++++
>   3 files changed, 26 insertions(+), 16 deletions(-)
>
> diff --git a/drivers/gpu/drm/drm_gpuvm.c b/drivers/gpu/drm/drm_gpuvm.c
> index 08c088319652..d7367a202fee 100644
> --- a/drivers/gpu/drm/drm_gpuvm.c
> +++ b/drivers/gpu/drm/drm_gpuvm.c
> @@ -614,12 +614,12 @@ static int __drm_gpuva_insert(struct drm_gpuvm *gpuvm,
>   static void __drm_gpuva_remove(struct drm_gpuva *va);
>   
>   static bool
> -drm_gpuvm_check_overflow(u64 addr, u64 range)
> +drm_gpuvm_check_overflow(struct drm_gpuvm *gpuvm, u64 addr, u64 range)
>   {
>   	u64 end;
>   
> -	return WARN(check_add_overflow(addr, range, &end),
> -		    "GPUVA address limited to %zu bytes.\n", sizeof(end));
> +	return drm_WARN(gpuvm->drm, check_add_overflow(addr, range, &end),
> +			"GPUVA address limited to %zu bytes.\n", sizeof(end));
>   }
>   
>   static bool
> @@ -647,7 +647,7 @@ static bool
>   drm_gpuvm_range_valid(struct drm_gpuvm *gpuvm,
>   		      u64 addr, u64 range)
>   {
> -	return !drm_gpuvm_check_overflow(addr, range) &&
> +	return !drm_gpuvm_check_overflow(gpuvm, addr, range) &&
>   	       drm_gpuvm_in_mm_range(gpuvm, addr, range) &&
>   	       !drm_gpuvm_in_kernel_node(gpuvm, addr, range);

When those parameters come from userspace you don't really want a 
warning in the system log in the first place.

Otherwise userspace can trivially spam the system log with warnings. The 
usual approach is to make this debug level severity instead.

Regards,
Christian.

>   }
> @@ -656,6 +656,7 @@ drm_gpuvm_range_valid(struct drm_gpuvm *gpuvm,
>    * drm_gpuvm_init() - initialize a &drm_gpuvm
>    * @gpuvm: pointer to the &drm_gpuvm to initialize
>    * @name: the name of the GPU VA space
> + * @drm: the &drm_device this VM resides in
>    * @start_offset: the start offset of the GPU VA space
>    * @range: the size of the GPU VA space
>    * @reserve_offset: the start of the kernel reserved GPU VA area
> @@ -668,8 +669,8 @@ drm_gpuvm_range_valid(struct drm_gpuvm *gpuvm,
>    * &name is expected to be managed by the surrounding driver structures.
>    */
>   void
> -drm_gpuvm_init(struct drm_gpuvm *gpuvm,
> -	       const char *name,
> +drm_gpuvm_init(struct drm_gpuvm *gpuvm, const char *name,
> +	       struct drm_device *drm,
>   	       u64 start_offset, u64 range,
>   	       u64 reserve_offset, u64 reserve_range,
>   	       const struct drm_gpuvm_ops *ops)
> @@ -677,20 +678,20 @@ drm_gpuvm_init(struct drm_gpuvm *gpuvm,
>   	gpuvm->rb.tree = RB_ROOT_CACHED;
>   	INIT_LIST_HEAD(&gpuvm->rb.list);
>   
> -	drm_gpuvm_check_overflow(start_offset, range);
> -	gpuvm->mm_start = start_offset;
> -	gpuvm->mm_range = range;
> -
>   	gpuvm->name = name ? name : "unknown";
>   	gpuvm->ops = ops;
> +	gpuvm->drm = drm;
>   
> -	memset(&gpuvm->kernel_alloc_node, 0, sizeof(struct drm_gpuva));
> +	drm_gpuvm_check_overflow(gpuvm, start_offset, range);
> +	gpuvm->mm_start = start_offset;
> +	gpuvm->mm_range = range;
>   
> +	memset(&gpuvm->kernel_alloc_node, 0, sizeof(struct drm_gpuva));
>   	if (reserve_range) {
>   		gpuvm->kernel_alloc_node.va.addr = reserve_offset;
>   		gpuvm->kernel_alloc_node.va.range = reserve_range;
>   
> -		if (likely(!drm_gpuvm_check_overflow(reserve_offset,
> +		if (likely(!drm_gpuvm_check_overflow(gpuvm, reserve_offset,
>   						     reserve_range)))
>   			__drm_gpuva_insert(gpuvm, &gpuvm->kernel_alloc_node);
>   	}
> @@ -712,8 +713,8 @@ drm_gpuvm_destroy(struct drm_gpuvm *gpuvm)
>   	if (gpuvm->kernel_alloc_node.va.range)
>   		__drm_gpuva_remove(&gpuvm->kernel_alloc_node);
>   
> -	WARN(!RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root),
> -	     "GPUVA tree is not empty, potentially leaking memory.");
> +	drm_WARN(gpuvm->drm, !RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root),
> +		 "GPUVA tree is not empty, potentially leaking memory.\n");
>   }
>   EXPORT_SYMBOL_GPL(drm_gpuvm_destroy);
>   
> @@ -795,7 +796,8 @@ drm_gpuva_remove(struct drm_gpuva *va)
>   	struct drm_gpuvm *gpuvm = va->vm;
>   
>   	if (unlikely(va == &gpuvm->kernel_alloc_node)) {
> -		WARN(1, "Can't destroy kernel reserved node.\n");
> +		drm_WARN(gpuvm->drm, 1,
> +			 "Can't destroy kernel reserved node.\n");
>   		return;
>   	}
>   
> diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.c b/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> index 5cf892c50f43..aaf5d28bd587 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.c
> @@ -1808,6 +1808,7 @@ int
>   nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli,
>   		  u64 kernel_managed_addr, u64 kernel_managed_size)
>   {
> +	struct drm_device *drm = cli->drm->dev;
>   	int ret;
>   	u64 kernel_managed_end = kernel_managed_addr + kernel_managed_size;
>   
> @@ -1836,7 +1837,7 @@ nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli,
>   	uvmm->kernel_managed_addr = kernel_managed_addr;
>   	uvmm->kernel_managed_size = kernel_managed_size;
>   
> -	drm_gpuvm_init(&uvmm->base, cli->name,
> +	drm_gpuvm_init(&uvmm->base, cli->name, drm,
>   		       NOUVEAU_VA_SPACE_START,
>   		       NOUVEAU_VA_SPACE_END,
>   		       kernel_managed_addr, kernel_managed_size,
> diff --git a/include/drm/drm_gpuvm.h b/include/drm/drm_gpuvm.h
> index bdfafc4a7705..687fd5893624 100644
> --- a/include/drm/drm_gpuvm.h
> +++ b/include/drm/drm_gpuvm.h
> @@ -29,6 +29,7 @@
>   #include <linux/rbtree.h>
>   #include <linux/types.h>
>   
> +#include <drm/drm_device.h>
>   #include <drm/drm_gem.h>
>   
>   struct drm_gpuvm;
> @@ -201,6 +202,11 @@ struct drm_gpuvm {
>   	 */
>   	const char *name;
>   
> +	/**
> +	 * @drm: the &drm_device this VM lives in
> +	 */
> +	struct drm_device *drm;
> +
>   	/**
>   	 * @mm_start: start of the VA space
>   	 */
> @@ -241,6 +247,7 @@ struct drm_gpuvm {
>   };
>   
>   void drm_gpuvm_init(struct drm_gpuvm *gpuvm, const char *name,
> +		    struct drm_device *drm,
>   		    u64 start_offset, u64 range,
>   		    u64 reserve_offset, u64 reserve_range,
>   		    const struct drm_gpuvm_ops *ops);

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ