[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20140914160752.1cb404cc@bbrezillon>
Date: Sun, 14 Sep 2014 16:07:52 +0200
From: Boris BREZILLON <boris.brezillon@...e-electrons.com>
To: Brian Norris <computersforpeace@...il.com>
Cc: Huang Shijie <shijie8@...il.com>,
Huang Shijie <shijie.huang@...el.com>,
Mike Voytovich <mvoytovich@...pal.com>,
linux-kernel@...r.kernel.org, Huang Shijie <b32955@...escale.com>,
linux-mtd@...ts.infradead.org, Roy Lee <roylee@...pal.com>,
David Woodhouse <dwmw2@...radead.org>,
linux-arm-kernel@...ts.infradead.org
Subject: Re: [PATCH] mtd: nand: gpmi: add proper raw access support
Hi Brian, Huang,
On Sat, 13 Sep 2014 10:38:41 -0700
Brian Norris <computersforpeace@...il.com> wrote:
> On Sat, Sep 13, 2014 at 11:36:24PM +0800, Huang Shijie wrote:
> > On Fri, Sep 12, 2014 at 02:30:50PM +0200, Boris BREZILLON wrote:
> > > This test validates what's returned by ecc_strength file in sysfs
> > > (which in turn is specified by the NAND controller when initializing
> > > the NAND chip).
> > >
> > > Doing this should not imply knowing the ECC algorithm in use in the
> > > NAND controller or the layout used to store data on NAND.
> > the difficulty is that the ECC parity area can be not byte aligned.
>
> Is there a problem with just rounding up to the nearest byte alignment
> and ignoring the few bits that are wasted?
>
> > As I ever said, it is hard to implement the two hooks.
>
> "Hard" doesn't mean we shouldn't. I really would like to encourage more
> NAND drivers to be programmed against the expected MTD behavior -- that
> (if possible with the given hardware) they can pass the MTD tests
> (drivers/mtd/tests/*).
Here is a draft for a gpmi_move_bits function we could use to move bits
(not bytes :-) from one memory region to another:
void gpmi_move_bits(u8 *dst, size_t dst_bit_off,
const u8 *src, size_t src_bit_off,
size_t nbits)
{
size_t i;
size_t nbytes;
u32 src_byte = 0;
src += src_bit_off / 8;
src_bit_off %= 8;
dst += dst_bit_off / 8;
dst_bit_off %= 8;
if (src_bit_off) {
src_byte = src[0] >> src_bit_off;
nbits -= 8 - src_bit_off;
src++;
}
nbytes = nbits / 8;
if (src_bit_off <= dst_bit_off) {
dst[0] &= GENMASK(dst_bit_off - 1, 0);
dst[0] |= src_byte << dst_bit_off;
src_bit_off += (8 - dst_bit_off);
src_byte >>= (8 - dst_bit_off);
dst_bit_off = 0;
dst++;
} else if (nbytes) {
src_byte |= src[0] << (8 - src_bit_off);
dst[0] &= GENMASK(dst_bit_off - 1, 0);
dst[0] |= src_byte << dst_bit_off;
src_bit_off += dst_bit_off;
src_byte >>= (8 - dst_bit_off);
dst_bit_off = 0;
dst++;
nbytes--;
src++;
if (src_bit_off > 7) {
src_bit_off -= 8;
dst[0] = src_byte;
dst++;
src_byte >>= 8;
}
}
if (!src_bit_off && !dst_bit_off) {
if (nbytes)
memcpy(dst, src, nbytes);
} else {
for (i = 0; i < nbytes; i++) {
src_byte |= src[i] << (8 - src_bit_off);
dst[i] = src_byte;
src_byte >>= 8;
}
}
dst += nbytes;
src += nbytes;
nbits %= 8;
if (!nbits && !src_bit_off)
return;
if (nbits)
src_byte |= (*src & GENMASK(nbits - 1, 0)) <<
((8 - src_bit_off) % 8);
nbits += (8 - src_bit_off) % 8;
if (dst_bit_off)
src_byte = (src_byte << dst_bit_off) |
(*dst & GENMASK(dst_bit_off - 1, 0));
nbits += dst_bit_off;
if (nbits % 8)
src_byte |= (dst[nbits / 8] & GENMASK(7, nbits % 8)) <<
(nbits / 8);
nbytes = DIV_ROUND_UP(nbits, 8);
for (i = 0; i < nbytes; i++) {
dst[i] = src_byte;
src_byte >>= 8;
}
}
I haven't tested it, and I think there is room for optimization.
My point is that performance is not a key aspect of raw functions
(those are often used by testing and debugging tools), hence we could
rely on this move_bits function to address the ECC bit alignment
problem.
Let me know what's your opinion on this approach.
Best Regards,
Boris
--
Boris Brezillon, Free Electrons
Embedded Linux and Kernel engineering
http://free-electrons.com
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists