lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <20180429111605.4cdf5ab2@bbrezillon> Date: Sun, 29 Apr 2018 11:16:05 +0200 From: Boris Brezillon <boris.brezillon@...tlin.com> To: Kees Cook <keescook@...omium.org> Cc: linux-kernel@...r.kernel.org, David Woodhouse <dwmw2@...radead.org>, Brian Norris <computersforpeace@...il.com>, Marek Vasut <marek.vasut@...il.com>, Richard Weinberger <richard@....at>, linux-mtd@...ts.infradead.org Subject: Re: [PATCH] mtd: nftl: Remove VLA usage Hi Kees, On Mon, 23 Apr 2018 13:35:00 -0700 Kees Cook <keescook@...omium.org> wrote: > On the quest to remove all VLAs from the kernel[1] this changes the > check_free_sectors() routine to use the same stack buffer for both > data byte checks (SECTORSIZE) and oob byte checks (oobsize). Since > these regions aren't needed at the same time, they don't need to be > consecutively allocated. Additionally, while it's possible for oobsize > to be large, it is unlikely to be larger than the actual SECTORSIZE. As > such, remove the VLA, adjust offsets and add a sanity check to make sure > we never get a pathological oobsize. > > [1] https://lkml.kernel.org/r/CA+55aFzCG-zNmZwX4A2FQpadafLfEzK6CC=qPXydAacU1RqZWA@mail.gmail.com > > Signed-off-by: Kees Cook <keescook@...omium.org> > --- > drivers/mtd/inftlmount.c | 11 ++++++++--- > drivers/mtd/nftlmount.c | 11 ++++++++--- > 2 files changed, 16 insertions(+), 6 deletions(-) > > diff --git a/drivers/mtd/inftlmount.c b/drivers/mtd/inftlmount.c > index aab4f68bd36f..9cdae7f0fc2e 100644 > --- a/drivers/mtd/inftlmount.c > +++ b/drivers/mtd/inftlmount.c > @@ -334,7 +334,7 @@ static int memcmpb(void *a, int c, int n) > static int check_free_sectors(struct INFTLrecord *inftl, unsigned int address, > int len, int check_oob) > { > - u8 buf[SECTORSIZE + inftl->mbd.mtd->oobsize]; > + u8 buf[SECTORSIZE]; Could we instead move to dynamic allocation. I mean, SECTORSIZE is 512 bytes, so only with this function we consume 1/16 of the stack. Not to mention that some MTD drivers might want to do DMA on buffer passed by the MTD user, and if the buffer is on the stack they'll have to use a bounce buffer instead. Regards, Boris > struct mtd_info *mtd = inftl->mbd.mtd; > size_t retlen; > int i; > @@ -346,10 +346,15 @@ static int check_free_sectors(struct INFTLrecord *inftl, unsigned int address, > return -1; > > if (check_oob) { > + if (mtd->oobsize > sizeof(buf)) { > + pr_warn("MTD oobsize > SECTORSIZE: %d\n", > + mtd->oobsize); > + return -1; > + } > if(inftl_read_oob(mtd, address, mtd->oobsize, > - &retlen, &buf[SECTORSIZE]) < 0) > + &retlen, buf) < 0) > return -1; > - if (memcmpb(buf + SECTORSIZE, 0xff, mtd->oobsize) != 0) > + if (memcmpb(buf, 0xff, mtd->oobsize) != 0) > return -1; > } > address += SECTORSIZE; > diff --git a/drivers/mtd/nftlmount.c b/drivers/mtd/nftlmount.c > index a6fbfa4e5799..e6eba7f3fdf5 100644 > --- a/drivers/mtd/nftlmount.c > +++ b/drivers/mtd/nftlmount.c > @@ -272,7 +272,7 @@ static int memcmpb(void *a, int c, int n) > static int check_free_sectors(struct NFTLrecord *nftl, unsigned int address, int len, > int check_oob) > { > - u8 buf[SECTORSIZE + nftl->mbd.mtd->oobsize]; > + u8 buf[SECTORSIZE]; > struct mtd_info *mtd = nftl->mbd.mtd; > size_t retlen; > int i; > @@ -284,10 +284,15 @@ static int check_free_sectors(struct NFTLrecord *nftl, unsigned int address, int > return -1; > > if (check_oob) { > + if (mtd->oobsize > sizeof(buf)) { > + pr_warn("MTD oobsize > SECTORSIZE: %d\n", > + mtd->oobsize); > + return -1; > + } > if(nftl_read_oob(mtd, address, mtd->oobsize, > - &retlen, &buf[SECTORSIZE]) < 0) > + &retlen, buf) < 0) > return -1; > - if (memcmpb(buf + SECTORSIZE, 0xff, mtd->oobsize) != 0) > + if (memcmpb(buf, 0xff, mtd->oobsize) != 0) > return -1; > } > address += SECTORSIZE;
Powered by blists - more mailing lists