[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20090915114546.3663f4a3.kamezawa.hiroyu@jp.fujitsu.com>
Date: Tue, 15 Sep 2009 11:45:46 +0900
From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
To: Wu Fengguang <fengguang.wu@...el.com>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
Benjamin Herrenschmidt <benh@...nel.crashing.org>,
Greg KH <greg@...ah.com>, Andi Kleen <andi@...stfloor.org>,
Christoph Lameter <clameter@....com>,
Ingo Molnar <mingo@...e.hu>, Tejun Heo <tj@...nel.org>,
Nick Piggin <npiggin@...e.de>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 3/3] HWPOISON: prevent /dev/kmem users from accessing
hwpoison pages
On Tue, 15 Sep 2009 10:18:54 +0800
Wu Fengguang <fengguang.wu@...el.com> wrote:
> When /dev/kmem read()/write() encounters hwpoison page, stop it
> and return the amount of work done till now.
>
Reviewed-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
> CC: Greg KH <greg@...ah.com>
> CC: Andi Kleen <andi@...stfloor.org>
> CC: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
> Signed-off-by: Wu Fengguang <fengguang.wu@...el.com>
> ---
> drivers/char/mem.c | 18 +++++++++++----
> mm/vmalloc.c | 51 +++++++++++++++++++++++++++----------------
> 2 files changed, 47 insertions(+), 22 deletions(-)
>
> --- linux-mm.orig/drivers/char/mem.c 2009-09-15 10:14:20.000000000 +0800
> +++ linux-mm/drivers/char/mem.c 2009-09-15 10:14:25.000000000 +0800
> @@ -429,6 +429,9 @@ static ssize_t read_kmem(struct file *fi
> */
> kbuf = xlate_dev_kmem_ptr((char *)p);
>
> + if (unlikely(virt_addr_valid(kbuf) &&
> + PageHWPoison(virt_to_page(kbuf))))
> + return -EIO;
> if (copy_to_user(buf, kbuf, sz))
> return -EFAULT;
> buf += sz;
> @@ -474,6 +477,7 @@ do_write_kmem(unsigned long p, const cha
> {
> ssize_t written, sz;
> unsigned long copied;
> + int err = 0;
>
> written = 0;
> #ifdef __ARCH_HAS_NO_PAGE_ZERO_MAPPED
> @@ -500,13 +504,19 @@ do_write_kmem(unsigned long p, const cha
> */
> ptr = xlate_dev_kmem_ptr((char *)p);
>
> + if (unlikely(virt_addr_valid(ptr) &&
> + PageHWPoison(virt_to_page(ptr)))) {
> + err = -EIO;
> + break;
> + }
> +
> copied = copy_from_user(ptr, buf, sz);
> if (copied) {
> written += sz - copied;
> - if (written)
> - break;
> - return -EFAULT;
> + err = -EFAULT;
> + break;
> }
> +
> buf += sz;
> p += sz;
> count -= sz;
> @@ -514,7 +524,7 @@ do_write_kmem(unsigned long p, const cha
> }
>
> *ppos += written;
> - return written;
> + return written ? written : err;
> }
>
>
> --- linux-mm.orig/mm/vmalloc.c 2009-09-15 10:14:18.000000000 +0800
> +++ linux-mm/mm/vmalloc.c 2009-09-15 10:17:20.000000000 +0800
> @@ -1661,6 +1661,8 @@ static int aligned_vread(char *buf, char
> if (length > count)
> length = count;
> p = vmalloc_to_page(addr);
> + if (unlikely(p && PageHWPoison(p)))
> + break;
> /*
> * To do safe access to this _mapped_ area, we need
> * lock. But adding lock here means that we need to add
> @@ -1700,6 +1702,8 @@ static int aligned_vwrite(char *buf, cha
> if (length > count)
> length = count;
> p = vmalloc_to_page(addr);
> + if (unlikely(p && PageHWPoison(p)))
> + break;
> /*
> * To do safe access to this _mapped_ area, we need
> * lock. But adding lock here means that we need to add
> @@ -1731,8 +1735,10 @@ static int aligned_vwrite(char *buf, cha
> * @count: number of bytes to be read.
> *
> * Returns # of bytes which addr and buf should be increased.
> - * (same number to @count). Returns 0 if [addr...addr+count) doesn't
> - * includes any intersect with alive vmalloc area.
> + * (same number to @count if no hwpoison pages encountered).
> + *
> + * Returns 0 if [addr...addr+count) doesn't includes any intersect with
> + * alive vmalloc area.
> *
> * This function checks that addr is a valid vmalloc'ed area, and
> * copy data from that area to a given buffer. If the given memory range
> @@ -1740,8 +1746,6 @@ static int aligned_vwrite(char *buf, cha
> * proper area of @buf. If there are memory holes, they'll be zero-filled.
> * IOREMAP area is treated as memory hole and no copy is done.
> *
> - * If [addr...addr+count) doesn't includes any intersects with alive
> - * vm_struct area, returns 0.
> * @buf should be kernel's buffer. Because this function uses KM_USER0,
> * the caller should guarantee KM_USER0 is not used.
> *
> @@ -1757,7 +1761,8 @@ long vread(char *buf, char *addr, unsign
> struct vm_struct *tmp;
> char *vaddr, *buf_start = buf;
> unsigned long buflen = count;
> - unsigned long n;
> + unsigned long ret = 0;
> + unsigned long n = 0;
>
> /* Don't allow overflow */
> if ((unsigned long) addr + count < count)
> @@ -1780,12 +1785,16 @@ long vread(char *buf, char *addr, unsign
> if (n > count)
> n = count;
> if (!(tmp->flags & VM_IOREMAP))
> - aligned_vread(buf, addr, n);
> - else /* IOREMAP area is treated as memory hole */
> + ret = aligned_vread(buf, addr, n);
> + else { /* IOREMAP area is treated as memory hole */
> memset(buf, 0, n);
> - buf += n;
> - addr += n;
> - count -= n;
> + ret = n;
> + }
> + buf += ret;
> + addr += ret;
> + count -= ret;
> + if (ret < n)
> + break;
> }
> finished:
> read_unlock(&vmlist_lock);
> @@ -1796,7 +1805,7 @@ finished:
> if (buf != buf_start + buflen)
> memset(buf, 0, buflen - (buf - buf_start));
>
> - return buflen;
> + return ret == n ? buflen : buflen - count;
> }
>
> /**
> @@ -1806,7 +1815,7 @@ finished:
> * @count: number of bytes to be read.
> *
> * Returns # of bytes which addr and buf should be increased.
> - * (same number to @count).
> + * (same number to @count if no hwpoison pages encountered).
> *
> * This function checks that addr is a valid vmalloc'ed area, and
> * copy data from a buffer to the given addr. If specified range of
> @@ -1829,7 +1838,9 @@ long vwrite(char *buf, char *addr, unsig
> {
> struct vm_struct *tmp;
> char *vaddr;
> - unsigned long n, buflen;
> + unsigned long buflen;
> + unsigned long ret = 0;
> + unsigned long n = 0;
>
> /* Don't allow overflow */
> if ((unsigned long) addr + count < count)
> @@ -1852,14 +1863,18 @@ long vwrite(char *buf, char *addr, unsig
> if (n > count)
> n = count;
> if (!(tmp->flags & VM_IOREMAP))
> - aligned_vwrite(buf, addr, n);
> - buf += n;
> - addr += n;
> - count -= n;
> + ret = aligned_vwrite(buf, addr, n);
> + else
> + ret = n;
> + buf += ret;
> + addr += ret;
> + count -= ret;
> + if (ret < n)
> + break;
> }
> finished:
> read_unlock(&vmlist_lock);
> - return buflen;
> + return ret == n ? buflen : buflen - count;
> }
>
> /**
>
> --
>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists