lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110429011113.GA3391@Krystal>
Date:	Thu, 28 Apr 2011 21:11:13 -0400
From:	Mathieu Desnoyers <mathieu.desnoyers@...icios.com>
To:	Huang Ying <ying.huang@...el.com>
Cc:	Len Brown <lenb@...nel.org>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	Andi Kleen <andi@...stfloor.org>,
	"Luck, Tony" <tony.luck@...el.com>,
	"linux-acpi@...r.kernel.org" <linux-acpi@...r.kernel.org>,
	Andrew Morton <akpm@...ux-foundation.org>
Subject: Re: [PATCH -v4 3/4] lib, Make gen_pool memory allocator lockless

* Huang Ying (ying.huang@...el.com) wrote:
> Hi, Mathieu,
> 
> Thanks for your comments.
> 
> On 04/28/2011 10:37 PM, Mathieu Desnoyers wrote:
> > * Huang Ying (ying.huang@...el.com) wrote:
> [snip]
> >>
> >> +/**
> >> + * gen_pool_for_each_chunk - iterate over chunks of generic memory pool
> >> + * @chunk:   the struct gen_pool_chunk * to use as a loop cursor
> >> + * @pool:    the generic memory pool
> >> + *
> >> + * Not lockless, proper mutual exclusion is needed to use this macro
> >> + * with other gen_pool function simultaneously.
> >> + */
> >> +#define gen_pool_for_each_chunk(chunk, pool)                 \
> >> +     list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk)
> > 
> > Is it just me or this macro is never used ? Maybe you should consider
> > removing it.
> 
> This macro is not used in this patch.  But it is used in 4/4 of the
> patchset to free the backing pages before destroy the pool.

Depending on how frequently you want to use it, you might want to use
list_for_each_entry_rcu directly rather than a macro wrapper. E.g.  for
2-3 uses, adding a macro just obfuscates the code IMHO (e.g. you don't
know it iterates on a RCU list by looking at the caller code).

> 
> [snip]
> >> @@ -108,43 +226,50 @@ EXPORT_SYMBOL(gen_pool_destroy);
> >>   * @size: number of bytes to allocate from the pool
> >>   *
> >>   * Allocate the requested number of bytes from the specified pool.
> >> - * Uses a first-fit algorithm.
> >> + * Uses a first-fit algorithm. Can not be used in NMI handler on
> >> + * architectures without NMI-safe cmpxchg implementation.
> >>   */
> >>  unsigned long gen_pool_alloc(struct gen_pool *pool, size_t size)
> >>  {
> >> -     struct list_head *_chunk;
> >>       struct gen_pool_chunk *chunk;
> >> -     unsigned long addr, flags;
> >> +     unsigned long addr;
> >>       int order = pool->min_alloc_order;
> >> -     int nbits, start_bit, end_bit;
> >> +     int nbits, start_bit = 0, end_bit, remain;
> >> +
> >> +#ifndef CONFIG_ARCH_HAVE_NMI_SAFE_CMPXCHG
> >> +     BUG_ON(in_nmi());
> >> +#endif
> >>
> >>       if (size == 0)
> >>               return 0;
> >>
> >>       nbits = (size + (1UL << order) - 1) >> order;
> >> -
> >> -     read_lock(&pool->lock);
> >> -     list_for_each(_chunk, &pool->chunks) {
> >> -             chunk = list_entry(_chunk, struct gen_pool_chunk, next_chunk);
> >> +     rcu_read_lock();
> >> +     list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) {
> >> +             if (size > atomic_read(&chunk->avail))
> >> +                     continue;
> >>
> >>               end_bit = (chunk->end_addr - chunk->start_addr) >> order;
> >> -
> >> -             spin_lock_irqsave(&chunk->lock, flags);
> >> -             start_bit = bitmap_find_next_zero_area(chunk->bits, end_bit, 0,
> >> -                                             nbits, 0);
> >> -             if (start_bit >= end_bit) {
> >> -                     spin_unlock_irqrestore(&chunk->lock, flags);
> >> +retry:
> >> +             start_bit = bitmap_find_next_zero_area(chunk->bits, end_bit,
> >> +                                                    start_bit, nbits, 0);
> >> +             if (start_bit >= end_bit)
> >>                       continue;
> >> +             remain = bitmap_set_ll(chunk->bits, start_bit, nbits);
> >> +             if (remain) {
> >> +                     remain = bitmap_clear_ll(chunk->bits, start_bit,
> >> +                                              nbits - remain);
> >> +                     BUG_ON(remain);
> >> +                     goto retry;
> >>               }
> >>
> >>               addr = chunk->start_addr + ((unsigned long)start_bit << order);
> >> -
> >> -             bitmap_set(chunk->bits, start_bit, nbits);
> >> -             spin_unlock_irqrestore(&chunk->lock, flags);
> >> -             read_unlock(&pool->lock);
> >> +             size = nbits << order;
> >> +             atomic_sub(size, &chunk->avail);
> >> +             rcu_read_unlock();
> > 
> > I don't really like seeing a rcu_read_unlock() within a rcu list
> > iteration (even if it comes right before a "return"). Doing:
> > 
> > unsigned long addr = 0;
> > 
> > rcu_read_lock();
> > list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) {
> >   if (...)
> >     continue;
> >   ...
> >   addr = ...;
> >   break;
> > }
> > rcu_read_unlock();
> > return addr;
> > 
> > Would be more symmetric, and would remove one return path, which makes
> > the code easier to modify in the future.
> 
> Unlock in loop is common in Linux kernel.  Sometimes it makes code
> cleaner (but not always).  Yes, for this case, we can avoid unlock in
> loop easily.  But for the next case it is not so clean.

See comment below,

> 
> >>               return addr;
> >>       }
> >> -     read_unlock(&pool->lock);
> >> +     rcu_read_unlock();
> >>       return 0;
> >>  }
> >>  EXPORT_SYMBOL(gen_pool_alloc);
> >> @@ -155,33 +280,73 @@ EXPORT_SYMBOL(gen_pool_alloc);
> >>   * @addr: starting address of memory to free back to pool
> >>   * @size: size in bytes of memory to free
> >>   *
> >> - * Free previously allocated special memory back to the specified pool.
> >> + * Free previously allocated special memory back to the specified
> >> + * pool.  Can not be used in NMI handler on architectures without
> >> + * NMI-safe cmpxchg implementation.
> >>   */
> >>  void gen_pool_free(struct gen_pool *pool, unsigned long addr, size_t size)
> >>  {
> >> -     struct list_head *_chunk;
> >>       struct gen_pool_chunk *chunk;
> >> -     unsigned long flags;
> >>       int order = pool->min_alloc_order;
> >> -     int bit, nbits;
> >> +     int start_bit, nbits, remain;
> >>
> >> -     nbits = (size + (1UL << order) - 1) >> order;
> >> -
> >> -     read_lock(&pool->lock);
> >> -     list_for_each(_chunk, &pool->chunks) {
> >> -             chunk = list_entry(_chunk, struct gen_pool_chunk, next_chunk);
> >> +#ifndef CONFIG_ARCH_HAVE_NMI_SAFE_CMPXCHG
> >> +     BUG_ON(in_nmi());
> >> +#endif
> >>
> >> +     nbits = (size + (1UL << order) - 1) >> order;

you could add:

  remain = nbits;

> >> +     rcu_read_lock();
> >> +     list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) {
> >>               if (addr >= chunk->start_addr && addr < chunk->end_addr) {
> >>                       BUG_ON(addr + size > chunk->end_addr);
> >> -                     spin_lock_irqsave(&chunk->lock, flags);
> >> -                     bit = (addr - chunk->start_addr) >> order;
> >> -                     while (nbits--)
> >> -                             __clear_bit(bit++, chunk->bits);
> >> -                     spin_unlock_irqrestore(&chunk->lock, flags);
> >> -                     break;
> >> +                     start_bit = (addr - chunk->start_addr) >> order;

You could turn this:

> >> +                     remain = bitmap_clear_ll(chunk->bits, start_bit, nbits);
> >> +                     BUG_ON(remain);
> >> +                     size = nbits << order;
> >> +                     atomic_add(size, &chunk->avail);

into:

  remain = bitmap_clear_ll(chunk->bits, start_bit, nbits);
  size = nbits << order;
  atomic_add(size, &chunk->avail);
  break;
    

> >> +                     rcu_read_unlock();
> > 
> > Same comment as above apply here.
> 
> It is harder to remove unlock in loop here.  An extra variable should be
> used to indicate that something is freed from the pool.  Do you think it
> is cleaner to just keep the unlock in loop here?
> 
> Best Regards,
> Huang Ying
> 
> > +                     return;
> >               }
> >       }

And turn this:

> > -     BUG_ON(nbits > 0);
> > -     read_unlock(&pool->lock);
> > +     rcu_read_unlock();
> > +     BUG();

into:

  BUG_ON(remain);
  rcu_read_unlock();

Does that look OK to you ? On the plus side, you end up having a single
BUG_ON() in the function.

Thanks,

Mathieu

> >  }
> [snip]

-- 
Mathieu Desnoyers
Operating System Efficiency R&D Consultant
EfficiOS Inc.
http://www.efficios.com
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ