[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <3fb178d8-f069-0ae2-1ed3-4ded84a71951@arm.com>
Date: Fri, 27 Sep 2019 10:55:33 +0100
From: Steven Price <steven.price@....com>
To: Neil Armstrong <narmstrong@...libre.com>, daniel@...ll.ch,
airlied@...ux.ie,
Christian König <christian.koenig@....com>
Cc: Tomeu Vizoso <tomeu.vizoso@...labora.com>,
linux-kernel@...r.kernel.org, dri-devel@...ts.freedesktop.org,
"open list:ARM/Amlogic Meson..." <linux-amlogic@...ts.infradead.org>,
Erico Nunes <nunes.erico@...il.com>,
Rob Herring <robh@...nel.org>
Subject: Re: drm_sched with panfrost crash on T820
On 27/09/2019 09:12, Neil Armstrong wrote:
> Hi Christian,
>
> In v5.3, running dEQP triggers the following kernel crash :
>
> [ 20.224982] Unable to handle kernel NULL pointer dereference at virtual address 0000000000000038
> [...]
> [ 20.291064] Hardware name: Khadas VIM2 (DT)
> [ 20.295217] Workqueue: events drm_sched_job_timedout
> [...]
> [ 20.304867] pc : drm_sched_increase_karma+0x5c/0xf0
> [ 20.309696] lr : drm_sched_increase_karma+0x44/0xf0
> [...]
> [ 20.396720] Call trace:
> [ 20.399138] drm_sched_increase_karma+0x5c/0xf0
> [ 20.403623] panfrost_job_timedout+0x12c/0x1e0
> [ 20.408021] drm_sched_job_timedout+0x48/0xa0
> [ 20.412336] process_one_work+0x1e0/0x320
> [ 20.416300] worker_thread+0x40/0x450
> [ 20.419924] kthread+0x124/0x128
> [ 20.423116] ret_from_fork+0x10/0x18
> [ 20.426653] Code: f9400001 540001c0 f9400a83 f9402402 (f9401c64)
> [ 20.432690] ---[ end trace bd02f890139096a7 ]---
>
> Which never happens, at all, on v5.2.
>
> I did a (very) long (7 days, ~100runs) bisect run using our LAVA lab (thanks tomeu !), but
> bisecting was not easy since the bad commit landed on drm-misc-next after v5.1-rc6, and
> then v5.2-rc1 was backmerged into drm-misc-next at:
> [1] 374ed5429346 Merge drm/drm-next into drm-misc-next
>
> Thus bisecting between [1] ang v5.2-rc1 leads to commit based on v5.2-rc1... where panfrost was
> not enabled in the Khadas VIM2 DT.
>
> Anyway, I managed to identify 3 possibly breaking commits :
> [2] 290764af7e36 drm/sched: Keep s_fence->parent pointer
> [3] 5918045c4ed4 drm/scheduler: rework job destruction
> [4] a5343b8a2ca5 drm/scheduler: Add flag to hint the release of guilty job.
My suspicion is that this is really a bug in Panfrost than the scheduler.
I can see the following sequence:
1. Jobs submitted to *both* slot 0 and slot 1 on the hardware.
2. Hardware 'hangs' (most likely an unhandled page fault)
3. The reset timeouts occur. Because Panfrost uses a separate scheduler
for each slot this means we have two threads racing to kill the
respective jobs.
4. Each thread calls drm_sched_job_timedout which calls
panfrost_job_timedout().
5. Both threads calling panfrost_job_timedout() execute:
for (i = 0; i < NUM_JOB_SLOTS; i++)
drm_sched_stop(&pfdev->js->queue[i].sched, sched_job);
6. drm_sched_stop() will free the job from the other scheduler. So both
jobs get freed (and sched->free_guilty is set).
7. Depending on the exact race either drm_sched_increase_karma()
dereferences the freed job, or drm_sched_job_timedout() attempts to
double-free the job (because free_guilty is set).
Now there is a lock (pfdev->reset_lock) held during steps 5-6, but the
problem is that the job which has timed out can be freed by the other
scheduler's timeout before/after that critical section.
One obvious issue with the DRM scheduler is that there is a call to
cancel_delayed_work() in drm_sched_stop() which to me looks like it
should be cancel_delayed_work_sync() to ensure that the timeout handling
has completed.
However in the above scenario a _sync() variety would then cause a
deadlock (one thread has pfdev->reset_lock and is waiting for the other
thread which is trying to take the lock).
So we need to update Panfrost so that it can coordinate the reset
between schedulers. Can you try something like the following (untested):
----8<---
diff --git a/drivers/gpu/drm/panfrost/panfrost_device.h b/drivers/gpu/drm/panfrost/panfrost_device.h
index f503c566e99f..6441c7fba6c4 100644
--- a/drivers/gpu/drm/panfrost/panfrost_device.h
+++ b/drivers/gpu/drm/panfrost/panfrost_device.h
@@ -99,6 +99,8 @@ struct panfrost_device {
unsigned long cur_volt;
struct panfrost_devfreq_slot slot[NUM_JOB_SLOTS];
} devfreq;
+
+ bool is_resetting;
};
struct panfrost_mmu {
diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c
index 05c85f45a0de..8b935680c066 100644
--- a/drivers/gpu/drm/panfrost/panfrost_job.c
+++ b/drivers/gpu/drm/panfrost/panfrost_job.c
@@ -388,6 +388,10 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job)
mutex_lock(&pfdev->reset_lock);
+ if (pfdev->is_resetting)
+ goto out;
+ pfdev->is_resetting = true;
+
for (i = 0; i < NUM_JOB_SLOTS; i++)
drm_sched_stop(&pfdev->js->queue[i].sched, sched_job);
@@ -406,6 +410,8 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job)
for (i = 0; i < NUM_JOB_SLOTS; i++)
drm_sched_start(&pfdev->js->queue[i].sched, true);
+ pfdev->is_resetting = false;
+out:
mutex_unlock(&pfdev->reset_lock);
}
Powered by blists - more mailing lists