lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Date:	Mon, 14 Dec 2009 14:39:20 +1100
From:	Stephen Rothwell <sfr@...b.auug.org.au>
To:	Jiri Slaby <jirislaby@...il.com>
Cc:	linux-next@...r.kernel.org, linux-kernel@...r.kernel.org,
	Al Viro <viro@...iv.linux.org.uk>
Subject: linux-next: manual merge of the limits tree with Linus' tree

Hi Jiri,

Today's linux-next merge of the limits tree got a conflict in mm/mremap.c
between commit 54f5de709984bae0d31d823ff03de755f9dcac54 ("untangling
do_mremap(), part 1") from Linus' tree and commit
518b0eb18cb50901611d5365d657e11ff0b4931f ("MM: use helpers for rlimits")
from the limits tree.

The former moved the code changed in the latter.  I fixed it up (see
below) and can carry the fix for a while.  Again, a merge with Linus'
tree will fix this up.
-- 
Cheers,
Stephen Rothwell                    sfr@...b.auug.org.au

diff --cc mm/mremap.c
index 8451908,d6740a1..0000000
--- a/mm/mremap.c
+++ b/mm/mremap.c
@@@ -261,137 -261,6 +261,137 @@@ static unsigned long move_vma(struct vm
  	return new_addr;
  }
  
 +static struct vm_area_struct *vma_to_resize(unsigned long addr,
 +	unsigned long old_len, unsigned long new_len, unsigned long *p)
 +{
 +	struct mm_struct *mm = current->mm;
 +	struct vm_area_struct *vma = find_vma(mm, addr);
 +
 +	if (!vma || vma->vm_start > addr)
 +		goto Efault;
 +
 +	if (is_vm_hugetlb_page(vma))
 +		goto Einval;
 +
 +	/* We can't remap across vm area boundaries */
 +	if (old_len > vma->vm_end - addr)
 +		goto Efault;
 +
 +	if (vma->vm_flags & (VM_DONTEXPAND | VM_PFNMAP)) {
 +		if (new_len > old_len)
 +			goto Efault;
 +	}
 +
 +	if (vma->vm_flags & VM_LOCKED) {
 +		unsigned long locked, lock_limit;
 +		locked = mm->locked_vm << PAGE_SHIFT;
- 		lock_limit = current->signal->rlim[RLIMIT_MEMLOCK].rlim_cur;
++		lock_limit = rlimit(RLIMIT_MEMLOCK);
 +		locked += new_len - old_len;
 +		if (locked > lock_limit && !capable(CAP_IPC_LOCK))
 +			goto Eagain;
 +	}
 +
 +	if (!may_expand_vm(mm, (new_len - old_len) >> PAGE_SHIFT))
 +		goto Enomem;
 +
 +	if (vma->vm_flags & VM_ACCOUNT) {
 +		unsigned long charged = (new_len - old_len) >> PAGE_SHIFT;
 +		if (security_vm_enough_memory(charged))
 +			goto Efault;
 +		*p = charged;
 +	}
 +
 +	return vma;
 +
 +Efault:	/* very odd choice for most of the cases, but... */
 +	return ERR_PTR(-EFAULT);
 +Einval:
 +	return ERR_PTR(-EINVAL);
 +Enomem:
 +	return ERR_PTR(-ENOMEM);
 +Eagain:
 +	return ERR_PTR(-EAGAIN);
 +}
 +
 +static unsigned long mremap_to(unsigned long addr,
 +	unsigned long old_len, unsigned long new_addr,
 +	unsigned long new_len)
 +{
 +	struct mm_struct *mm = current->mm;
 +	struct vm_area_struct *vma;
 +	unsigned long ret = -EINVAL;
 +	unsigned long charged = 0;
 +	unsigned long map_flags;
 +
 +	if (new_addr & ~PAGE_MASK)
 +		goto out;
 +
 +	if (new_len > TASK_SIZE || new_addr > TASK_SIZE - new_len)
 +		goto out;
 +
 +	/* Check if the location we're moving into overlaps the
 +	 * old location at all, and fail if it does.
 +	 */
 +	if ((new_addr <= addr) && (new_addr+new_len) > addr)
 +		goto out;
 +
 +	if ((addr <= new_addr) && (addr+old_len) > new_addr)
 +		goto out;
 +
 +	ret = security_file_mmap(NULL, 0, 0, 0, new_addr, 1);
 +	if (ret)
 +		goto out;
 +
 +	ret = do_munmap(mm, new_addr, new_len);
 +	if (ret)
 +		goto out;
 +
 +	if (old_len >= new_len) {
 +		ret = do_munmap(mm, addr+new_len, old_len - new_len);
 +		if (ret && old_len != new_len)
 +			goto out;
 +		old_len = new_len;
 +	}
 +
 +	vma = vma_to_resize(addr, old_len, new_len, &charged);
 +	if (IS_ERR(vma)) {
 +		ret = PTR_ERR(vma);
 +		goto out;
 +	}
 +
 +	map_flags = MAP_FIXED;
 +	if (vma->vm_flags & VM_MAYSHARE)
 +		map_flags |= MAP_SHARED;
 +
 +	ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff +
 +				((addr - vma->vm_start) >> PAGE_SHIFT),
 +				map_flags);
 +	if (ret & ~PAGE_MASK)
 +		goto out1;
 +
 +	ret = move_vma(vma, addr, old_len, new_len, new_addr);
 +	if (!(ret & ~PAGE_MASK))
 +		goto out;
 +out1:
 +	vm_unacct_memory(charged);
 +
 +out:
 +	return ret;
 +}
 +
 +static int vma_expandable(struct vm_area_struct *vma, unsigned long delta)
 +{
 +	unsigned long end = vma->vm_end + delta;
 +	if (end < vma->vm_end) /* overflow */
 +		return 0;
 +	if (vma->vm_next && vma->vm_next->vm_start < end) /* intersection */
 +		return 0;
 +	if (get_unmapped_area(NULL, vma->vm_start, end - vma->vm_start,
 +			      0, MAP_FIXED) & ~PAGE_MASK)
 +		return 0;
 +	return 1;
 +}
 +
  /*
   * Expand (or shrink) an existing mapping, potentially moving it at the
   * same time (controlled by the MREMAP_MAYMOVE flag and available VM space)
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ