[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <7d36c4a4-9446-4e45-e851-cc0d6b9f5dcb@intel.com>
Date: Thu, 20 Jan 2022 16:25:24 +0000
From: Matthew Auld <matthew.auld@...el.com>
To: Robert Beckett <bob.beckett@...labora.com>,
Ramalingam C <ramalingam.c@...el.com>
Cc: Jani Nikula <jani.nikula@...ux.intel.com>,
Joonas Lahtinen <joonas.lahtinen@...ux.intel.com>,
Rodrigo Vivi <rodrigo.vivi@...el.com>,
Tvrtko Ursulin <tvrtko.ursulin@...ux.intel.com>,
David Airlie <airlied@...ux.ie>,
Daniel Vetter <daniel@...ll.ch>,
intel-gfx@...ts.freedesktop.org, dri-devel@...ts.freedesktop.org,
linux-kernel@...r.kernel.org
Subject: Re: [PATCH v2 1/4] drm/i915: enforce min GTT alignment for discrete
cards
On 20/01/2022 16:09, Robert Beckett wrote:
>
>
> On 20/01/2022 15:58, Matthew Auld wrote:
>> On 20/01/2022 15:44, Robert Beckett wrote:
>>>
>>>
>>> On 20/01/2022 14:59, Matthew Auld wrote:
>>>> On 20/01/2022 13:15, Robert Beckett wrote:
>>>>>
>>>>>
>>>>> On 20/01/2022 11:46, Ramalingam C wrote:
>>>>>> On 2022-01-18 at 17:50:34 +0000, Robert Beckett wrote:
>>>>>>> From: Matthew Auld <matthew.auld@...el.com>
>>>>>>>
>>>>>>> For local-memory objects we need to align the GTT addresses
>>>>>>> to 64K, both for the ppgtt and ggtt.
>>>>>>>
>>>>>>> We need to support vm->min_alignment > 4K, depending
>>>>>>> on the vm itself and the type of object we are inserting.
>>>>>>> With this in mind update the GTT selftests to take this
>>>>>>> into account.
>>>>>>>
>>>>>>> For DG2 we further align and pad lmem object GTT addresses
>>>>>>> to 2MB to ensure PDEs contain consistent page sizes as
>>>>>>> required by the HW.
>>>>>>>
>>>>>>> Signed-off-by: Matthew Auld <matthew.auld@...el.com>
>>>>>>> Signed-off-by: Ramalingam C <ramalingam.c@...el.com>
>>>>>>> Signed-off-by: Robert Beckett <bob.beckett@...labora.com>
>>>>>>> Cc: Joonas Lahtinen <joonas.lahtinen@...ux.intel.com>
>>>>>>> Cc: Rodrigo Vivi <rodrigo.vivi@...el.com>
>>>>>>> ---
>>>>>>> .../i915/gem/selftests/i915_gem_client_blt.c | 23 +++--
>>>>>>> drivers/gpu/drm/i915/gt/intel_gtt.c | 14 +++
>>>>>>> drivers/gpu/drm/i915/gt/intel_gtt.h | 9 ++
>>>>>>> drivers/gpu/drm/i915/i915_vma.c | 14 +++
>>>>>>> drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 96
>>>>>>> ++++++++++++-------
>>>>>>> 5 files changed, 115 insertions(+), 41 deletions(-)
>>>>>>>
>>>>>>> diff --git
>>>>>>> a/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
>>>>>>> b/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
>>>>>>> index c08f766e6e15..7fee95a65414 100644
>>>>>>> --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
>>>>>>> +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
>>>>>>> @@ -39,6 +39,7 @@ struct tiled_blits {
>>>>>>> struct blit_buffer scratch;
>>>>>>> struct i915_vma *batch;
>>>>>>> u64 hole;
>>>>>>> + u64 align;
>>>>>>> u32 width;
>>>>>>> u32 height;
>>>>>>> };
>>>>>>> @@ -410,14 +411,21 @@ tiled_blits_create(struct intel_engine_cs
>>>>>>> *engine, struct rnd_state *prng)
>>>>>>> goto err_free;
>>>>>>> }
>>>>>>> - hole_size = 2 * PAGE_ALIGN(WIDTH * HEIGHT * 4);
>>>>>>> + t->align = I915_GTT_PAGE_SIZE_2M; /* XXX worst case, derive
>>>>>>> from vm! */
>>>>>>> + t->align = max(t->align,
>>>>>>> + i915_vm_min_alignment(t->ce->vm,
>>>>>>> INTEL_MEMORY_LOCAL));
>>>>>>> + t->align = max(t->align,
>>>>>>> + i915_vm_min_alignment(t->ce->vm,
>>>>>>> INTEL_MEMORY_SYSTEM));
>>>>>>> +
>>>>>>> + hole_size = 2 * round_up(WIDTH * HEIGHT * 4, t->align);
>>>>>>> hole_size *= 2; /* room to maneuver */
>>>>>>> - hole_size += 2 * I915_GTT_MIN_ALIGNMENT;
>>>>>>> + hole_size += 2 * t->align; /* padding on either side */
>>>>>>> mutex_lock(&t->ce->vm->mutex);
>>>>>>> memset(&hole, 0, sizeof(hole));
>>>>>>> err = drm_mm_insert_node_in_range(&t->ce->vm->mm, &hole,
>>>>>>> - hole_size, 0, I915_COLOR_UNEVICTABLE,
>>>>>>> + hole_size, t->align,
>>>>>>> + I915_COLOR_UNEVICTABLE,
>>>>>>> 0, U64_MAX,
>>>>>>> DRM_MM_INSERT_BEST);
>>>>>>> if (!err)
>>>>>>> @@ -428,7 +436,7 @@ tiled_blits_create(struct intel_engine_cs
>>>>>>> *engine, struct rnd_state *prng)
>>>>>>> goto err_put;
>>>>>>> }
>>>>>>> - t->hole = hole.start + I915_GTT_MIN_ALIGNMENT;
>>>>>>> + t->hole = hole.start + t->align;
>>>>>>> pr_info("Using hole at %llx\n", t->hole);
>>>>>>> err = tiled_blits_create_buffers(t, WIDTH, HEIGHT, prng);
>>>>>>> @@ -455,7 +463,7 @@ static void tiled_blits_destroy(struct
>>>>>>> tiled_blits *t)
>>>>>>> static int tiled_blits_prepare(struct tiled_blits *t,
>>>>>>> struct rnd_state *prng)
>>>>>>> {
>>>>>>> - u64 offset = PAGE_ALIGN(t->width * t->height * 4);
>>>>>>> + u64 offset = round_up(t->width * t->height * 4, t->align);
>>>>>>> u32 *map;
>>>>>>> int err;
>>>>>>> int i;
>>>>>>> @@ -486,8 +494,7 @@ static int tiled_blits_prepare(struct
>>>>>>> tiled_blits *t,
>>>>>>> static int tiled_blits_bounce(struct tiled_blits *t, struct
>>>>>>> rnd_state *prng)
>>>>>>> {
>>>>>>> - u64 offset =
>>>>>>> - round_up(t->width * t->height * 4, 2 *
>>>>>>> I915_GTT_MIN_ALIGNMENT);
>>>>>>> + u64 offset = round_up(t->width * t->height * 4, 2 * t->align);
>>>>>>> int err;
>>>>>>> /* We want to check position invariant tiling across GTT
>>>>>>> eviction */
>>>>>>> @@ -500,7 +507,7 @@ static int tiled_blits_bounce(struct
>>>>>>> tiled_blits *t, struct rnd_state *prng)
>>>>>>> /* Reposition so that we overlap the old addresses, and
>>>>>>> slightly off */
>>>>>>> err = tiled_blit(t,
>>>>>>> - &t->buffers[2], t->hole + I915_GTT_MIN_ALIGNMENT,
>>>>>>> + &t->buffers[2], t->hole + t->align,
>>>>>>> &t->buffers[1], t->hole + 3 * offset / 2);
>>>>>>> if (err)
>>>>>>> return err;
>>>>>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.c
>>>>>>> b/drivers/gpu/drm/i915/gt/intel_gtt.c
>>>>>>> index 46be4197b93f..7c92b25c0f26 100644
>>>>>>> --- a/drivers/gpu/drm/i915/gt/intel_gtt.c
>>>>>>> +++ b/drivers/gpu/drm/i915/gt/intel_gtt.c
>>>>>>> @@ -223,6 +223,20 @@ void i915_address_space_init(struct
>>>>>>> i915_address_space *vm, int subclass)
>>>>>>> GEM_BUG_ON(!vm->total);
>>>>>>> drm_mm_init(&vm->mm, 0, vm->total);
>>>>>>> +
>>>>>>> + memset64(vm->min_alignment, I915_GTT_MIN_ALIGNMENT,
>>>>>>> + ARRAY_SIZE(vm->min_alignment));
>>>>>>> +
>>>>>>> + if (HAS_64K_PAGES(vm->i915)) {
>>>>>>> + if (IS_DG2(vm->i915)) {
>>>>>> I think we need this 2M alignment for all platform with
>>>>>> HAS_64K_PAGES.
>>>>>> Not only for DG2.
>>>>>
>>>>> really? can we get confirmation of this?
>>>>> this contradicts the documentation in patch 4, which you reviewed,
>>>>> so I am confused now
>>>>
>>>> Starting from DG2, some platforms will have this new 64K GTT page
>>>> size restriction when dealing with LMEM. The HAS_64K_PAGES() macro
>>>> is meant to cover exactly that, AFAIK.
>>>
>>> As I understood it, 64K pages only are a requirement going forward
>>> for discrete cards, but the restriction of nt sharing pdes with 4k
>>> and 64k pages was specific to DG2.
>>>
>>> e.g. xehpsdv is also defined as having 64k pages. And others in
>>> future are likely to, but without the PDE sharing restrictions.
>>
>> Yeah, pretty much. But there is one other platform lurking.
>>
>> From chatting with Ram, it might also make sense to disentangle
>> HAS_64K_PAGES(), since it currently means both that we need min 64K
>> page granularity, and that there is this compact-pt layout thing which
>> doesn't allow mixing 64K and 4K in the same page-table.
>
> okay, so it sounds to me like the IS_DG2 check here is appropriate.
> Other 64K page systems will not have the 2MB alignment requirement.
There is both dg2 and xehpsdv as per i915_pci.c. IIRC xehpsdv came
first, and then dg2 inherited this feature. For example the accelerated
DG2 moves series[1] is meant to work on both platforms.
[1] https://patchwork.freedesktop.org/series/97544/
>
> If any future platform does require compact-pt layout, when adding that
> plaform, we can then add a HAS_COMPACT_PT macro or something, which
> would be set for DG2 and the future platform.
>
> For now, this code seems correct to me as it currently only affects DG2.
>>
>>>
>>> If this is not the case, and all 64K page devices will also
>>> necessitate not sharing PDEs, then we can just use the HAS_64K_PAGES
>>> and use 2MB everywhere, but so far this sounds unconfirmed.
>>>
>>>>
>>>>>
>>>>>>> + vm->min_alignment[INTEL_MEMORY_LOCAL] =
>>>>>>> I915_GTT_PAGE_SIZE_2M;
>>>>>>> + vm->min_alignment[INTEL_MEMORY_STOLEN_LOCAL] =
>>>>>>> I915_GTT_PAGE_SIZE_2M;
>>>>>>> + } else {
>>>>>>> + vm->min_alignment[INTEL_MEMORY_LOCAL] =
>>>>>>> I915_GTT_PAGE_SIZE_64K;
>>>>>>> + vm->min_alignment[INTEL_MEMORY_STOLEN_LOCAL] =
>>>>>>> I915_GTT_PAGE_SIZE_64K;
>>>>>>> + }
>>>>>>> + }
>>>>>>> +
>>>>>>> vm->mm.head_node.color = I915_COLOR_UNEVICTABLE;
>>>>>>> INIT_LIST_HEAD(&vm->bound_list);
>>>>>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>>>>> b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>>>>> index 8073438b67c8..b8da2514d601 100644
>>>>>>> --- a/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>>>>> +++ b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>>>>> @@ -29,6 +29,8 @@
>>>>>>> #include "i915_selftest.h"
>>>>>>> #include "i915_vma_resource.h"
>>>>>>> #include "i915_vma_types.h"
>>>>>>> +#include "i915_params.h"
>>>>>>> +#include "intel_memory_region.h"
>>>>>>> #define I915_GFP_ALLOW_FAIL (GFP_KERNEL | __GFP_RETRY_MAYFAIL |
>>>>>>> __GFP_NOWARN)
>>>>>>> @@ -223,6 +225,7 @@ struct i915_address_space {
>>>>>>> struct device *dma;
>>>>>>> u64 total; /* size addr space maps (ex. 2GB for
>>>>>>> ggtt) */
>>>>>>> u64 reserved; /* size addr space reserved */
>>>>>>> + u64 min_alignment[INTEL_MEMORY_STOLEN_LOCAL + 1];
>>>>>>> unsigned int bind_async_flags;
>>>>>>> @@ -384,6 +387,12 @@ i915_vm_has_scratch_64K(struct
>>>>>>> i915_address_space *vm)
>>>>>>> return vm->scratch_order == get_order(I915_GTT_PAGE_SIZE_64K);
>>>>>>> }
>>>>>>> +static inline u64 i915_vm_min_alignment(struct
>>>>>>> i915_address_space *vm,
>>>>>>> + enum intel_memory_type type)
>>>>>>> +{
>>>>>>> + return vm->min_alignment[type];
>>>>>>> +}
>>>>>>> +
>>>>>>> static inline bool
>>>>>>> i915_vm_has_cache_coloring(struct i915_address_space *vm)
>>>>>>> {
>>>>>>> diff --git a/drivers/gpu/drm/i915/i915_vma.c
>>>>>>> b/drivers/gpu/drm/i915/i915_vma.c
>>>>>>> index 1f15c3298112..9ac92e7a3566 100644
>>>>>>> --- a/drivers/gpu/drm/i915/i915_vma.c
>>>>>>> +++ b/drivers/gpu/drm/i915/i915_vma.c
>>>>>>> @@ -756,6 +756,20 @@ i915_vma_insert(struct i915_vma *vma, u64
>>>>>>> size, u64 alignment, u64 flags)
>>>>>>> }
>>>>>>> color = 0;
>>>>>>> +
>>>>>>> + if (HAS_64K_PAGES(vma->vm->i915) &&
>>>>>>> i915_gem_object_is_lmem(vma->obj)) {
>>>>>>> + alignment = max(alignment, I915_GTT_PAGE_SIZE_64K);
>>>>>>> + /*
>>>>>>> + * DG2 can not have different sized pages in any given
>>>>>>> PDE (2MB range).
>>>>>>> + * Keeping things simple, we force any lmem object to
>>>>>>> reserve
>>>>>>> + * 2MB chunks, preventing any smaller pages being used
>>>>>>> alongside
>>>>>>> + */
>>>>>>> + if (IS_DG2(vma->vm->i915)) {
>>>>>> Similarly here we dont need special case for DG2.
>>>>>>
>>>>>> Ram
>>>>>>> + alignment = max(alignment, I915_GTT_PAGE_SIZE_2M);
>>>>>>> + size = round_up(size, I915_GTT_PAGE_SIZE_2M);
>>>>>>> + }
>>>>>>> + }
>>>>>>> +
>>>>>>> if (i915_vm_has_cache_coloring(vma->vm))
>>>>>>> color = vma->obj->cache_level;
>>>>>>> diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
>>>>>>> b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
>>>>>>> index 076d860ce01a..2f3f0c01786b 100644
>>>>>>> --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
>>>>>>> +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
>>>>>>> @@ -238,6 +238,8 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> u64 hole_start, u64 hole_end,
>>>>>>> unsigned long end_time)
>>>>>>> {
>>>>>>> + const unsigned int min_alignment =
>>>>>>> + i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> I915_RND_STATE(seed_prng);
>>>>>>> struct i915_vma_resource *mock_vma_res;
>>>>>>> unsigned int size;
>>>>>>> @@ -251,9 +253,10 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> I915_RND_SUBSTATE(prng, seed_prng);
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> unsigned int *order, count, n;
>>>>>>> - u64 hole_size;
>>>>>>> + u64 hole_size, aligned_size;
>>>>>>> - hole_size = (hole_end - hole_start) >> size;
>>>>>>> + aligned_size = max_t(u32, ilog2(min_alignment), size);
>>>>>>> + hole_size = (hole_end - hole_start) >> aligned_size;
>>>>>>> if (hole_size > KMALLOC_MAX_SIZE / sizeof(u32))
>>>>>>> hole_size = KMALLOC_MAX_SIZE / sizeof(u32);
>>>>>>> count = hole_size >> 1;
>>>>>>> @@ -274,8 +277,8 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> GEM_BUG_ON(!order);
>>>>>>> - GEM_BUG_ON(count * BIT_ULL(size) > vm->total);
>>>>>>> - GEM_BUG_ON(hole_start + count * BIT_ULL(size) > hole_end);
>>>>>>> + GEM_BUG_ON(count * BIT_ULL(aligned_size) > vm->total);
>>>>>>> + GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) >
>>>>>>> hole_end);
>>>>>>> /* Ignore allocation failures (i.e. don't report them as
>>>>>>> * a test failure) as we are purposefully allocating very
>>>>>>> @@ -298,10 +301,10 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> for (n = 0; n < count; n++) {
>>>>>>> - u64 addr = hole_start + order[n] * BIT_ULL(size);
>>>>>>> + u64 addr = hole_start + order[n] *
>>>>>>> BIT_ULL(aligned_size);
>>>>>>> intel_wakeref_t wakeref;
>>>>>>> - GEM_BUG_ON(addr + BIT_ULL(size) > vm->total);
>>>>>>> + GEM_BUG_ON(addr + BIT_ULL(aligned_size) > vm->total);
>>>>>>> if (igt_timeout(end_time,
>>>>>>> "%s timed out before %d/%d\n",
>>>>>>> @@ -344,7 +347,7 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> mock_vma_res->bi.pages = obj->mm.pages;
>>>>>>> - mock_vma_res->node_size = BIT_ULL(size);
>>>>>>> + mock_vma_res->node_size = BIT_ULL(aligned_size);
>>>>>>> mock_vma_res->start = addr;
>>>>>>> with_intel_runtime_pm(vm->gt->uncore->rpm, wakeref)
>>>>>>> @@ -355,7 +358,7 @@ static int lowlevel_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> i915_random_reorder(order, count, &prng);
>>>>>>> for (n = 0; n < count; n++) {
>>>>>>> - u64 addr = hole_start + order[n] * BIT_ULL(size);
>>>>>>> + u64 addr = hole_start + order[n] *
>>>>>>> BIT_ULL(aligned_size);
>>>>>>> intel_wakeref_t wakeref;
>>>>>>> GEM_BUG_ON(addr + BIT_ULL(size) > vm->total);
>>>>>>> @@ -399,8 +402,10 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> {
>>>>>>> const u64 hole_size = hole_end - hole_start;
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> + const unsigned int min_alignment =
>>>>>>> + i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> const unsigned long max_pages =
>>>>>>> - min_t(u64, ULONG_MAX - 1, hole_size/2 >> PAGE_SHIFT);
>>>>>>> + min_t(u64, ULONG_MAX - 1, (hole_size / 2) >>
>>>>>>> ilog2(min_alignment));
>>>>>>> const unsigned long max_step = max(int_sqrt(max_pages), 2UL);
>>>>>>> unsigned long npages, prime, flags;
>>>>>>> struct i915_vma *vma;
>>>>>>> @@ -441,14 +446,17 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> offset = p->offset;
>>>>>>> list_for_each_entry(obj, &objects, st_link) {
>>>>>>> + u64 aligned_size = round_up(obj->base.size,
>>>>>>> + min_alignment);
>>>>>>> +
>>>>>>> vma = i915_vma_instance(obj, vm, NULL);
>>>>>>> if (IS_ERR(vma))
>>>>>>> continue;
>>>>>>> if (p->step < 0) {
>>>>>>> - if (offset < hole_start + obj->base.size)
>>>>>>> + if (offset < hole_start + aligned_size)
>>>>>>> break;
>>>>>>> - offset -= obj->base.size;
>>>>>>> + offset -= aligned_size;
>>>>>>> }
>>>>>>> err = i915_vma_pin(vma, 0, 0, offset | flags);
>>>>>>> @@ -470,22 +478,25 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> i915_vma_unpin(vma);
>>>>>>> if (p->step > 0) {
>>>>>>> - if (offset + obj->base.size > hole_end)
>>>>>>> + if (offset + aligned_size > hole_end)
>>>>>>> break;
>>>>>>> - offset += obj->base.size;
>>>>>>> + offset += aligned_size;
>>>>>>> }
>>>>>>> }
>>>>>>> offset = p->offset;
>>>>>>> list_for_each_entry(obj, &objects, st_link) {
>>>>>>> + u64 aligned_size = round_up(obj->base.size,
>>>>>>> + min_alignment);
>>>>>>> +
>>>>>>> vma = i915_vma_instance(obj, vm, NULL);
>>>>>>> if (IS_ERR(vma))
>>>>>>> continue;
>>>>>>> if (p->step < 0) {
>>>>>>> - if (offset < hole_start + obj->base.size)
>>>>>>> + if (offset < hole_start + aligned_size)
>>>>>>> break;
>>>>>>> - offset -= obj->base.size;
>>>>>>> + offset -= aligned_size;
>>>>>>> }
>>>>>>> if (!drm_mm_node_allocated(&vma->node) ||
>>>>>>> @@ -506,22 +517,25 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> if (p->step > 0) {
>>>>>>> - if (offset + obj->base.size > hole_end)
>>>>>>> + if (offset + aligned_size > hole_end)
>>>>>>> break;
>>>>>>> - offset += obj->base.size;
>>>>>>> + offset += aligned_size;
>>>>>>> }
>>>>>>> }
>>>>>>> offset = p->offset;
>>>>>>> list_for_each_entry_reverse(obj, &objects,
>>>>>>> st_link) {
>>>>>>> + u64 aligned_size = round_up(obj->base.size,
>>>>>>> + min_alignment);
>>>>>>> +
>>>>>>> vma = i915_vma_instance(obj, vm, NULL);
>>>>>>> if (IS_ERR(vma))
>>>>>>> continue;
>>>>>>> if (p->step < 0) {
>>>>>>> - if (offset < hole_start + obj->base.size)
>>>>>>> + if (offset < hole_start + aligned_size)
>>>>>>> break;
>>>>>>> - offset -= obj->base.size;
>>>>>>> + offset -= aligned_size;
>>>>>>> }
>>>>>>> err = i915_vma_pin(vma, 0, 0, offset | flags);
>>>>>>> @@ -543,22 +557,25 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> i915_vma_unpin(vma);
>>>>>>> if (p->step > 0) {
>>>>>>> - if (offset + obj->base.size > hole_end)
>>>>>>> + if (offset + aligned_size > hole_end)
>>>>>>> break;
>>>>>>> - offset += obj->base.size;
>>>>>>> + offset += aligned_size;
>>>>>>> }
>>>>>>> }
>>>>>>> offset = p->offset;
>>>>>>> list_for_each_entry_reverse(obj, &objects,
>>>>>>> st_link) {
>>>>>>> + u64 aligned_size = round_up(obj->base.size,
>>>>>>> + min_alignment);
>>>>>>> +
>>>>>>> vma = i915_vma_instance(obj, vm, NULL);
>>>>>>> if (IS_ERR(vma))
>>>>>>> continue;
>>>>>>> if (p->step < 0) {
>>>>>>> - if (offset < hole_start + obj->base.size)
>>>>>>> + if (offset < hole_start + aligned_size)
>>>>>>> break;
>>>>>>> - offset -= obj->base.size;
>>>>>>> + offset -= aligned_size;
>>>>>>> }
>>>>>>> if (!drm_mm_node_allocated(&vma->node) ||
>>>>>>> @@ -579,9 +596,9 @@ static int fill_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> if (p->step > 0) {
>>>>>>> - if (offset + obj->base.size > hole_end)
>>>>>>> + if (offset + aligned_size > hole_end)
>>>>>>> break;
>>>>>>> - offset += obj->base.size;
>>>>>>> + offset += aligned_size;
>>>>>>> }
>>>>>>> }
>>>>>>> }
>>>>>>> @@ -611,6 +628,7 @@ static int walk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> const u64 hole_size = hole_end - hole_start;
>>>>>>> const unsigned long max_pages =
>>>>>>> min_t(u64, ULONG_MAX - 1, hole_size >> PAGE_SHIFT);
>>>>>>> + unsigned long min_alignment;
>>>>>>> unsigned long flags;
>>>>>>> u64 size;
>>>>>>> @@ -620,6 +638,8 @@ static int walk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> if (i915_is_ggtt(vm))
>>>>>>> flags |= PIN_GLOBAL;
>>>>>>> + min_alignment = i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> +
>>>>>>> for_each_prime_number_from(size, 1, max_pages) {
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> struct i915_vma *vma;
>>>>>>> @@ -638,7 +658,7 @@ static int walk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> for (addr = hole_start;
>>>>>>> addr + obj->base.size < hole_end;
>>>>>>> - addr += obj->base.size) {
>>>>>>> + addr += round_up(obj->base.size, min_alignment)) {
>>>>>>> err = i915_vma_pin(vma, 0, 0, addr | flags);
>>>>>>> if (err) {
>>>>>>> pr_err("%s bind failed at %llx + %llx [hole
>>>>>>> %llx- %llx] with err=%d\n",
>>>>>>> @@ -690,6 +710,7 @@ static int pot_hole(struct i915_address_space
>>>>>>> *vm,
>>>>>>> {
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> struct i915_vma *vma;
>>>>>>> + unsigned int min_alignment;
>>>>>>> unsigned long flags;
>>>>>>> unsigned int pot;
>>>>>>> int err = 0;
>>>>>>> @@ -698,6 +719,8 @@ static int pot_hole(struct i915_address_space
>>>>>>> *vm,
>>>>>>> if (i915_is_ggtt(vm))
>>>>>>> flags |= PIN_GLOBAL;
>>>>>>> + min_alignment = i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> +
>>>>>>> obj = i915_gem_object_create_internal(vm->i915, 2 *
>>>>>>> I915_GTT_PAGE_SIZE);
>>>>>>> if (IS_ERR(obj))
>>>>>>> return PTR_ERR(obj);
>>>>>>> @@ -710,13 +733,13 @@ static int pot_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> /* Insert a pair of pages across every pot boundary within
>>>>>>> the hole */
>>>>>>> for (pot = fls64(hole_end - 1) - 1;
>>>>>>> - pot > ilog2(2 * I915_GTT_PAGE_SIZE);
>>>>>>> + pot > ilog2(2 * min_alignment);
>>>>>>> pot--) {
>>>>>>> u64 step = BIT_ULL(pot);
>>>>>>> u64 addr;
>>>>>>> - for (addr = round_up(hole_start + I915_GTT_PAGE_SIZE,
>>>>>>> step) - I915_GTT_PAGE_SIZE;
>>>>>>> - addr <= round_down(hole_end - 2*I915_GTT_PAGE_SIZE,
>>>>>>> step) - I915_GTT_PAGE_SIZE;
>>>>>>> + for (addr = round_up(hole_start + min_alignment, step) -
>>>>>>> min_alignment;
>>>>>>> + addr <= round_down(hole_end - (2 * min_alignment),
>>>>>>> step) - min_alignment;
>>>>>>> addr += step) {
>>>>>>> err = i915_vma_pin(vma, 0, 0, addr | flags);
>>>>>>> if (err) {
>>>>>>> @@ -761,6 +784,7 @@ static int drunk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> unsigned long end_time)
>>>>>>> {
>>>>>>> I915_RND_STATE(prng);
>>>>>>> + unsigned int min_alignment;
>>>>>>> unsigned int size;
>>>>>>> unsigned long flags;
>>>>>>> @@ -768,15 +792,18 @@ static int drunk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> if (i915_is_ggtt(vm))
>>>>>>> flags |= PIN_GLOBAL;
>>>>>>> + min_alignment = i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> +
>>>>>>> /* Keep creating larger objects until one cannot fit into
>>>>>>> the hole */
>>>>>>> for (size = 12; (hole_end - hole_start) >> size; size++) {
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> unsigned int *order, count, n;
>>>>>>> struct i915_vma *vma;
>>>>>>> - u64 hole_size;
>>>>>>> + u64 hole_size, aligned_size;
>>>>>>> int err = -ENODEV;
>>>>>>> - hole_size = (hole_end - hole_start) >> size;
>>>>>>> + aligned_size = max_t(u32, ilog2(min_alignment), size);
>>>>>>> + hole_size = (hole_end - hole_start) >> aligned_size;
>>>>>>> if (hole_size > KMALLOC_MAX_SIZE / sizeof(u32))
>>>>>>> hole_size = KMALLOC_MAX_SIZE / sizeof(u32);
>>>>>>> count = hole_size >> 1;
>>>>>>> @@ -816,7 +843,7 @@ static int drunk_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> GEM_BUG_ON(vma->size != BIT_ULL(size));
>>>>>>> for (n = 0; n < count; n++) {
>>>>>>> - u64 addr = hole_start + order[n] * BIT_ULL(size);
>>>>>>> + u64 addr = hole_start + order[n] *
>>>>>>> BIT_ULL(aligned_size);
>>>>>>> err = i915_vma_pin(vma, 0, 0, addr | flags);
>>>>>>> if (err) {
>>>>>>> @@ -868,11 +895,14 @@ static int __shrink_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> {
>>>>>>> struct drm_i915_gem_object *obj;
>>>>>>> unsigned long flags = PIN_OFFSET_FIXED | PIN_USER;
>>>>>>> + unsigned int min_alignment;
>>>>>>> unsigned int order = 12;
>>>>>>> LIST_HEAD(objects);
>>>>>>> int err = 0;
>>>>>>> u64 addr;
>>>>>>> + min_alignment = i915_vm_min_alignment(vm, INTEL_MEMORY_SYSTEM);
>>>>>>> +
>>>>>>> /* Keep creating larger objects until one cannot fit into
>>>>>>> the hole */
>>>>>>> for (addr = hole_start; addr < hole_end; ) {
>>>>>>> struct i915_vma *vma;
>>>>>>> @@ -913,7 +943,7 @@ static int __shrink_hole(struct
>>>>>>> i915_address_space *vm,
>>>>>>> }
>>>>>>> i915_vma_unpin(vma);
>>>>>>> - addr += size;
>>>>>>> + addr += round_up(size, min_alignment);
>>>>>>> /*
>>>>>>> * Since we are injecting allocation faults at random
>>>>>>> intervals,
>>>>>>> --
>>>>>>> 2.25.1
>>>>>>>
Powered by blists - more mailing lists