lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Sun, 29 Mar 2009 10:35:31 -0400
From:	Rik van Riel <riel@...hat.com>
To:	Martin Schwidefsky <schwidefsky@...ibm.com>
CC:	linux-mm@...ck.org, linux-kernel@...r.kernel.org,
	virtualization@...ts.osdl.org, frankeh@...son.ibm.com,
	akpm@...l.org, nickpiggin@...oo.com.au, hugh@...itas.com
Subject: Re: [patch 1/6] Guest page hinting: core + volatile page cache.

Martin Schwidefsky wrote:
> On Fri, 27 Mar 2009 18:57:31 -0400
> Rik van Riel <riel@...hat.com> wrote:
>> Martin Schwidefsky wrote:

>>>   There are some alternatives how this can be done, e.g. a global
>>>   lock, or lock per segment in the kernel page table, or the per page
>>>   bit PG_arch_1 if it is still free.
>> Can this be taken care of by memory barriers and
>> careful ordering of operations?
> 
> I don't see how this could be done with memory barries, the sequence is
> 1) check conditions
> 2) do state change to volatile
> 
> another cpus can do
> i) change one of the conditions
> 
> The operation i) needs to be postponed while the first cpu has done 1)
> but not done 2) yet. 1+2 needs to be atomic but consists of several
> instructions. Ergo we need a lock, no ?

You are right.

Hashed locks may be a space saving option, with a
set of (cache line aligned?) locks in each zone
and the page state lock chosen by taking a hash
of the page number or address.

Not ideal, but at least we can get some NUMA
locality.

>>> +	if (page->index != linear_page_index(vma, addr))
>>> +		/* If nonlinear, store the file page offset in the pte. */
>>> +		set_pte_at(dst_mm, addr, dst_pte, pgoff_to_pte(page->index));
>>> +	else
>>> +		pte_clear(dst_mm, addr, dst_pte);
>>>  }
>> It would be good to document that PG_discarded can only happen for
>> file pages and NOT for eg. clean swap cache pages.
> 
> PG_discarded can happen for swap cache pages as well. If a clean swap
> cache page gets remove and subsequently access again the discard fault
> handler will set the bit (see __page_discard). The code necessary for
> volatile swap cache is introduced with patch #2. So I would rather not
> add a comment in patch #1 only to remove it again with patch #2 ..

I discovered that once I opened the next email :)

>>> @@ -1390,6 +1391,7 @@ int test_clear_page_writeback(struct pag
>>>  			radix_tree_tag_clear(&mapping->page_tree,
>>>  						page_index(page),
>>>  						PAGECACHE_TAG_WRITEBACK);
>>> +			page_make_volatile(page, 1);
>>>  			if (bdi_cap_account_writeback(bdi)) {
>>>  				__dec_bdi_stat(bdi, BDI_WRITEBACK);
>>>  				__bdi_writeout_inc(bdi);
>> Does this mark the page volatile before the IO writing the
>> dirty data back to disk has even started?  Is that OK?
>  
> Hmm, it could be that the page_make_volatile is just superflouos here.
> The logic here is that whenever one of the conditions that prevent a
> page from becoming volatile is cleared a try with page_make_volatile
> is done. The condition in question here is PageWriteback(page). If we
> can prove that one of the other conditions is true this particular call
> is a waste of effort.

Actually, test_clear_page_writeback is probably called
on IO completion and it was just me being confused after
a few hundred lines of very new (to me) VM code :)

I guess the patch is correct.

Acked-by: Rik van Riel <riel@...hat.com>

-- 
All rights reversed.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ