[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87jzgj1ejc.fsf@mail.lhotse>
Date: Wed, 14 Aug 2024 22:20:55 +1000
From: Michael Ellerman <mpe@...erman.id.au>
To: Niklas Cassel <cassel@...nel.org>
Cc: Kolbjørn Barmen <linux-ppc@...la.no>,
linuxppc-dev@...ts.ozlabs.org,
linux-kernel@...r.kernel.org, linux-ide@...r.kernel.org, Jonáš Vidra
<vidra@...l.mff.cuni.cz>, Christoph Hellwig <hch@....de>,
linux@...ck-us.net
Subject: Re: Since 6.10 - kernel oops/panics on G4 macmini due to change in
drivers/ata/pata_macio.c
Niklas Cassel <cassel@...nel.org> writes:
> On Tue, Aug 13, 2024 at 10:32:36PM +1000, Michael Ellerman wrote:
>> Niklas Cassel <cassel@...nel.org> writes:
>> > On Tue, Aug 13, 2024 at 07:49:34AM +0200, Jonáš Vidra wrote:
...
>> >> ------------[ cut here ]------------
>> >> kernel BUG at drivers/ata/pata_macio.c:544!
>> >
>> > https://github.com/torvalds/linux/blob/v6.11-rc3/drivers/ata/pata_macio.c#L544
>> >
>> > It seems that the
>> > while (sg_len) loop does not play nice with the new .max_segment_size.
>>
>> Right, but only for 4KB kernels for some reason. Is there some limit
>> elsewhere that prevents the bug tripping on 64KB kernels, or is it just
>> luck that no one has hit it?
>
> Have your tried running fio (flexible I/O tester), with reads with a very
> large block sizes?
>
> I would be surprised if it isn't possible to trigger the same bug with
> 64K page size.
>
> max segment size = 64K
> MAX_DCMDS = 256
> 256 * 64K = 16 MiB
> What happens if you run fio with a 16 MiB blocksize?
>
> Something like:
> $ sudo fio --name=test --filename=/dev/sdX --direct=1 --runtime=60 --ioengine=io_uring --rw=read --iodepth=4 --bs=16M
Nothing interesting happens, fio succeeds.
The largest request that comes into pata_macio_qc_prep() is 1280KB,
which results in 40 DMA list entries.
I tried with a larger block size but it doesn't change anything. I guess
there's some limit somewhere else in the stack?
That was testing on qemu, but I don't think it should matter?
I guess there's no way to run the fio test against a file, ie. without a
raw partition? My real G5 doesn't have any spare disks/partitions in it.
cheers
fio-3.37
Starting 1 process
test: (groupid=0, jobs=1): err= 0: pid=257: Wed Aug 14 22:18:59 2024
read: IOPS=6, BW=195MiB/s (204MB/s)(96.0MiB/493msec)
slat (usec): min=32973, max=35222, avg=33836.35, stdev=1212.51
clat (msec): min=378, max=448, avg=413.35, stdev=34.99
lat (msec): min=413, max=481, avg=447.19, stdev=33.87
clat percentiles (msec):
| 1.00th=[ 380], 5.00th=[ 380], 10.00th=[ 380], 20.00th=[ 380],
| 30.00th=[ 380], 40.00th=[ 414], 50.00th=[ 414], 60.00th=[ 414],
| 70.00th=[ 447], 80.00th=[ 447], 90.00th=[ 447], 95.00th=[ 447],
| 99.00th=[ 447], 99.50th=[ 447], 99.90th=[ 447], 99.95th=[ 447],
| 99.99th=[ 447]
bw ( KiB/s): min=195047, max=195047, per=97.82%, avg=195047.00, stdev= 0.00, samples=1
iops : min= 5, max= 5, avg= 5.00, stdev= 0.00, samples=1
lat (msec) : 500=100.00%
cpu : usr=1.62%, sys=11.97%, ctx=22, majf=0, minf=1540
IO depths : 1=33.3%, 2=66.7%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued rwts: total=3,0,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=4
Run status group 0 (all jobs):
READ: bw=195MiB/s (204MB/s), 195MiB/s-195MiB/s (204MB/s-204MB/s), io=96.0MiB (101MB), run=493-493msec
Disk stats (read/write):
sda: ios=78/0, sectors=196608/0, merge=0/0, ticks=745/0, in_queue=745, util=66.89%
Powered by blists - more mailing lists