[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <E1HVZwW-0008SG-00@dorka.pomaz.szeredi.hu>
Date: Sun, 25 Mar 2007 23:08:00 +0200
From: Miklos Szeredi <miklos@...redi.hu>
To: a.p.zijlstra@...llo.nl
CC: akpm@...ux-foundation.org, linux-kernel@...r.kernel.org,
linux-mm@...ck.org
Subject: Re: [patch 3/3] update ctime and mtime for mmaped write
> A few comments..
Thanks for reviewing.
> > Index: linux-2.6.21-rc4-mm1/mm/rmap.c
> > ===================================================================
> > --- linux-2.6.21-rc4-mm1.orig/mm/rmap.c 2007-03-24 19:03:11.000000000 +0100
> > +++ linux-2.6.21-rc4-mm1/mm/rmap.c 2007-03-24 19:34:30.000000000 +0100
> > @@ -507,6 +507,43 @@ int page_mkclean(struct page *page)
> > EXPORT_SYMBOL_GPL(page_mkclean);
> >
> > /**
> > + * test_clear_page_modified - check and clear the dirty bit for all mappings of a page
> > + * @page: the page to check
> > + */
> > +bool test_clear_page_modified(struct page *page)
> > +{
> > + struct address_space *mapping = page->mapping;
>
> page_mapping(page)? Otherwise that BUG_ON(!mapping) a few lines down
> isn't of much use.
OK, removed BUG_ON(). This is called with page locked and mapping
checked, so there's no need to use page_mapping().
> > + spin_lock(&mapping->i_mmap_lock);
> > + vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) {
> > + if (vma->vm_flags & VM_SHARED) {
> > + struct mm_struct *mm = vma->vm_mm;
> > + unsigned long addr = vma_address(page, vma);
> > + pte_t *pte;
> > + spinlock_t *ptl;
> > +
> > + if (addr != -EFAULT &&
> > + (pte = page_check_address(page, mm, addr, &ptl))) {
> > + if (ptep_clear_flush_dirty(vma, addr, pte))
> > + modified = true;
> > + pte_unmap_unlock(pte, ptl);
> > + }
>
> Its against coding style to do assignments in conditionals.
OK, cleaned up.
> > + if (page_test_and_clear_dirty(page))
> > + modified = true;
> > + return modified;
> > +}
>
> Why not parametrize page_mkclean() to conditionally wrprotect clean
> pages? Something like:
Well, I don't really like this, because for msync, there's really no
need to do the complex ptep operations. Using ptep_clear_flush_dirty()
can save a couple of cycles.
> > + mapping = vma->vm_file->f_mapping;
> > + modified = test_and_clear_bit(AS_CMTIME, &mapping->flags);
> > +
> > + pagevec_init(&pvec, 0);
> > + index = linear_page_index(vma, start);
> > + end_index = linear_page_index(vma, end);
> > + while (index < end_index) {
> > + int i;
> > + int nr_pages = min(end_index - index, (pgoff_t) PAGEVEC_SIZE);
> > +
> > + if (mapping_cap_account_dirty(mapping))
> > + nr_pages = pagevec_lookup_tag(&pvec, mapping, &index,
> > + PAGECACHE_TAG_DIRTY, nr_pages);
> > + else
> > + nr_pages = pagevec_lookup(&pvec, mapping, index,
> > + nr_pages);
> > + if (!nr_pages)
> > + break;
> > +
> > + for (i = 0; i < nr_pages; i++) {
> > + struct page *page = pvec.pages[i];
> > +
> > + /* Skip pages which are just being read */
> > + if (!PageUptodate(page))
> > + continue;
> > +
> > + lock_page(page);
> > + index = page->index + 1;
> > + if (page->mapping == mapping &&
> > + test_clear_page_modified(page)) {
>
> page_mkclean(page, 0)
>
> > + set_page_dirty(page);
>
> set_page_dirty_mapping() ?
That would cause the file times to be updated twice, which is wrong.
But if AS_CMTIME were tested/cleared after walking the pages, this
would work. Fixed.
Also realized, that setting AS_CMTIME from the page fault can also
cause a double update, since the PTE dirty bit is also set there.
And also realized, that doing the file update from munmap can also
cause a double update, if this is not the last mapping of the inode.
Fixed all these in the updated patch.
Thanks,
Miklos
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists