[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <173382321353.8959.8314520413901294535@jlahtine-mobl.ger.corp.intel.com>
Date: Tue, 10 Dec 2024 11:33:33 +0200
From: Joonas Lahtinen <joonas.lahtinen@...ux.intel.com>
To: Andrzej Hajda <andrzej.hajda@...el.com>, Christian König <christian.koenig@....com>, Jonathan Cavitt <jonathan.cavitt@...el.com>, Linux MM <linux-mm@...ck.org>, Maciej Patelczyk <maciej.patelczyk@...el.com>, Mika Kuoppala <mika.kuoppala@...ux.intel.com>, dri-devel@...ts.freedesktop.org, intel-xe@...ts.freedesktop.org, lkml <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 14/26] drm/xe/eudebug: implement userptr_vma access
Quoting Christian König (2024-12-09 17:42:32)
> Am 09.12.24 um 16:31 schrieb Simona Vetter:
> > On Mon, Dec 09, 2024 at 03:03:04PM +0100, Christian König wrote:
> >> Am 09.12.24 um 14:33 schrieb Mika Kuoppala:
> >>> From: Andrzej Hajda <andrzej.hajda@...el.com>
> >>>
> >>> Debugger needs to read/write program's vmas including userptr_vma.
> >>> Since hmm_range_fault is used to pin userptr vmas, it is possible
> >>> to map those vmas from debugger context.
> >> Oh, this implementation is extremely questionable as well. Adding the LKML
> >> and the MM list as well.
> >>
> >> First of all hmm_range_fault() does *not* pin anything!
> >>
> >> In other words you don't have a page reference when the function returns,
> >> but rather just a sequence number you can check for modifications.
> > I think it's all there, holds the invalidation lock during the critical
> > access/section, drops it when reacquiring pages, retries until it works.
> >
> > I think the issue is more that everyone hand-rolls userptr.
>
> Well that is part of the issue.
>
> The general problem here is that the eudebug interface tries to simulate
> the memory accesses as they would have happened by the hardware.
Could you elaborate, what is that a problem in that, exactly?
It's pretty much the equivalent of ptrace() poke/peek but for GPU memory.
And it is exactly the kind of interface that makes sense for debugger as
GPU memory != CPU memory, and they don't need to align at all.
> What the debugger should probably do is to cleanly attach to the
> application, get the information which CPU address is mapped to which
> GPU address and then use the standard ptrace interfaces.
I don't quite agree here -- at all. "Which CPU address is mapped to
which GPU address" makes no sense when the GPU address space and CPU
address space is completely controlled by the userspace driver/application.
Please try to consider things outside of the ROCm architecture.
Something like a register scratch region or EU instructions should not
even be mapped to CPU address space as CPU has no business accessing it
during normal operation. And backing of such region will vary per
context/LRC on the same virtual address per EU thread.
You seem to be suggesting to rewrite even our userspace driver to behave
the same way as ROCm driver does just so that we could implement debug memory
accesses via ptrace() to the CPU address space.
That seems bit of a radical suggestion, especially given the drawbacks
pointed out in your suggested design.
> The whole interface re-invents a lot of functionality which is already
> there
I'm not really sure I would call adding a single interface for memory
reading and writing to be "re-inventing a lot of functionality".
All the functionality behind this interface will be needed by GPU core
dumping, anyway. Just like for the other patch series.
> just because you don't like the idea to attach to the debugged
> application in userspace.
A few points that have been brought up as drawback to the
GPU debug through ptrace(), but to recap a few relevant ones for this
discussion:
- You can only really support GDB stop-all mode or at least have to
stop all the CPU threads while you control the GPU threads to
avoid interference. Elaborated on this on the other threads more.
- Controlling the GPU threads will always interfere with CPU threads.
Doesn't seem feasible to single-step an EU thread while CPU threads
continue to run freely?
- You are very much restricted by the CPU VA ~ GPU VA alignment
requirement, which is not true for OpenGL or Vulkan etc. Seems
like one of the reasons why ROCm debugging is not easily extendable
outside compute?
- You have to expose extra memory to CPU process just for GPU
debugger access and keep track of GPU VA for each. Makes the GPU more
prone to OOB writes from CPU. Exactly what not mapping the memory
to CPU tried to protect the GPU from to begin with.
> As far as I can see this whole idea is extremely questionable. This
> looks like re-inventing the wheel in a different color.
I see it like reinventing a round wheel compared to octagonal wheel.
Could you elaborate with facts much more on your position why the ROCm
debugger design is an absolute must for others to adopt?
Otherwise it just looks like you are trying to prevent others from
implementing a more flexible debugging interface through vague comments about
"questionable design" without going into details. Not listing much concrete
benefits nor addressing the very concretely expressed drawbacks of your
suggested design, makes it seem like a very biased non-technical discussion.
So while review interest and any comments are very much appreciated, please
also work on providing bit more reasoning and facts instead of just claiming
things. That'll help make the discussion much more fruitful.
Regards, Joonas
Powered by blists - more mailing lists