[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <9ede7211-ae58-5cd4-4cf6-74c1f508f1a6@huawei.com>
Date: Tue, 10 May 2022 14:44:50 +0100
From: John Garry <john.garry@...wei.com>
To: Jens Axboe <axboe@...nel.dk>, <linux-block@...r.kernel.org>
CC: <linux-kernel@...r.kernel.org>, <linux-scsi@...r.kernel.org>
Subject: Re: [RFC PATCH 0/2] sbitmap: NUMA node spreading
On 10/05/2022 13:50, Jens Axboe wrote:
>> fio config:
>> bs=4096, iodepth=128, numjobs=10, cpus_allowed_policy=split, rw=read,
>> ioscheduler=none
>>
>> Before:
>> 7130K
>>
>> After:
>> 7630K
>>
>> So a +7% IOPS gain.
Thanks for having a look.
> What does the comparison run on a non-NUMA non-shared queue look like?
> Because I bet it'd be slower.
I could test more to get a solid result for that.
>
> To be honest, I don't like this approach at all. It makes the normal
> case quite a bit slower by having an extra layer of indirection for the
> word, that's quite a bit of extra cost.
Yes, there is the extra load. I would hope that there would be a low
cost, but I agree that we still want to avoid it. So prob no point in
testing this more.
> It doesn't seem like a good
> approach for the issue, as it pessimizes the normal fast case.
>
> Spreading the memory out does probably make sense, but we need to retain
> the fast normal case. Making sbitmap support both, selected at init
> time, would be far more likely to be acceptable imho.
I wanted to keep the code changes minimal for an initial RFC to test the
water.
My original approach did not introduce the extra load for normal path
and had some init time selection for a normal word map vs numa word map,
but the code grew and became somewhat unmanageable. I'll revisit it to
see how to improve that.
Cheers,
john
Powered by blists - more mailing lists