lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 13 Jul 2015 02:22:28 +0000
From:	Zhao Qiang <qiang.zhao@...escale.com>
To:	Laura Abbott <labbott@...hat.com>,
	"lauraa@...eaurora.org" <lauraa@...eaurora.org>
CC:	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	"linuxppc-dev@...ts.ozlabs.org" <linuxppc-dev@...ts.ozlabs.org>,
	"akpm@...ux-foundation.org" <akpm@...ux-foundation.org>,
	"olof@...om.net" <olof@...om.net>,
	"catalin.marinas@....com" <catalin.marinas@....com>,
	"Scott Wood" <scottwood@...escale.com>,
	Xiaobo Xie <X.Xie@...escale.com>
Subject: RE: [RFC] genalloc:add an gen_pool_alloc_align func to genalloc



> -----Original Message-----
> From: Laura Abbott [mailto:labbott@...hat.com]
> Sent: Friday, July 10, 2015 5:51 AM
> To: Zhao Qiang-B45475; lauraa@...eaurora.org
> Cc: linux-kernel@...r.kernel.org; linuxppc-dev@...ts.ozlabs.org;
> akpm@...ux-foundation.org; olof@...om.net; catalin.marinas@....com; Wood
> Scott-B07421; Xie Xiaobo-R63061
> Subject: Re: [RFC] genalloc:add an gen_pool_alloc_align func to genalloc
> 
> On 07/09/2015 12:47 AM, Zhao Qiang wrote:
> > Bytes alignment is required to manage some special ram, so add
> > gen_pool_alloc_align func to genalloc.
> > rename gen_pool_alloc to gen_pool_alloc_align with a align parameter,
> > then provide gen_pool_alloc to call gen_pool_alloc_align with align =
> > 1 Byte.
> >
> > Signed-off-by: Zhao Qiang <B45475@...escale.com>
> > ---
> > FSL's IP block QE require this function to manage muram.
> > QE supported only PowerPC, and its code was put under arch/powerpc
> > directory, using arch/powerpc/lib/rheap.c to manage muram.
> > Now it support both arm(ls1021,ls1043,ls2085 and such on) and powerpc,
> > the code need to move from arch/powerpc to public direcory, Scott wood
> > hopes to use genalloc to manage the muram, after discussing with
> > scott, we decide to add gen_pool_alloc_align to meet the requirement
> > for bytes-alignment.
> 
> gen_pool supports custom allocation algorithms. I thought this was
> discussed previously and the conclusion was that if you wanted alignment
> you should use custom allocation algorithms. I'm failing at finding any
> thread discussing it though.
> 
> Perhaps another option would be to add another runtime argument to
> gen_pool where you could pass the alignment to your custom allocation
> function. This way alignment isn't inherently coded into any of the
> algorithms.
> 
> >
> >   include/linux/genalloc.h | 10 +++++++---
> >   lib/genalloc.c           | 38 ++++++++++++++++++++++++++++++--------
> >   2 files changed, 37 insertions(+), 11 deletions(-)
> >
> > diff --git a/include/linux/genalloc.h b/include/linux/genalloc.h index
> > 1ccaab4..65fdf14 100644
> > --- a/include/linux/genalloc.h
> > +++ b/include/linux/genalloc.h
> > @@ -96,6 +96,8 @@ static inline int gen_pool_add(struct gen_pool *pool,
> unsigned long addr,
> >   }
> >   extern void gen_pool_destroy(struct gen_pool *);
> >   extern unsigned long gen_pool_alloc(struct gen_pool *, size_t);
> > +extern unsigned long gen_pool_alloc_align(struct gen_pool *, size_t,
> > +		unsigned long align);
> >   extern void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size,
> >   		dma_addr_t *dma);
> >   extern void gen_pool_free(struct gen_pool *, unsigned long, size_t);
> > @@ -108,14 +110,16 @@ extern void gen_pool_set_algo(struct gen_pool
> *pool, genpool_algo_t algo,
> >   		void *data);
> >
> >   extern unsigned long gen_pool_first_fit(unsigned long *map, unsigned
> long size,
> > -		unsigned long start, unsigned int nr, void *data);
> > +		unsigned long start, unsigned int nr, void *data,
> > +		unsigned long align_mask);
> >
> >   extern unsigned long gen_pool_first_fit_order_align(unsigned long
> *map,
> >   		unsigned long size, unsigned long start, unsigned int nr,
> > -		void *data);
> > +		void *data, unsigned long align_mask);
> >
> >   extern unsigned long gen_pool_best_fit(unsigned long *map, unsigned
> long size,
> > -		unsigned long start, unsigned int nr, void *data);
> > +		unsigned long start, unsigned int nr, void *data,
> > +		unsigned long align_mask);
> >
> >   extern struct gen_pool *devm_gen_pool_create(struct device *dev,
> >   		int min_alloc_order, int nid);
> > diff --git a/lib/genalloc.c b/lib/genalloc.c index d214866..dd63448
> > 100644
> > --- a/lib/genalloc.c
> > +++ b/lib/genalloc.c
> > @@ -258,19 +258,22 @@ void gen_pool_destroy(struct gen_pool *pool)
> >   EXPORT_SYMBOL(gen_pool_destroy);
> >
> >   /**
> > - * gen_pool_alloc - allocate special memory from the pool
> > + * gen_pool_alloc_align - allocate special memory from the pool
> >    * @pool: pool to allocate from
> >    * @size: number of bytes to allocate from the pool
> > + * @align: number of bytes to align
> >    *
> >    * Allocate the requested number of bytes from the specified pool.
> >    * Uses the pool allocation function (with first-fit algorithm by
> default).
> >    * Can not be used in NMI handler on architectures without
> >    * NMI-safe cmpxchg implementation.
> >    */
> > -unsigned long gen_pool_alloc(struct gen_pool *pool, size_t size)
> > +unsigned long gen_pool_alloc_align(struct gen_pool *pool, size_t size,
> > +		unsigned long align)
> >   {
> >   	struct gen_pool_chunk *chunk;
> >   	unsigned long addr = 0;
> > +	unsigned long align_mask;
> >   	int order = pool->min_alloc_order;
> >   	int nbits, start_bit = 0, end_bit, remain;
> >
> > @@ -281,6 +284,7 @@ unsigned long gen_pool_alloc(struct gen_pool *pool,
> size_t size)
> >   	if (size == 0)
> >   		return 0;
> >
> > +	align_mask = ((align + (1UL << order) - 1) >> order) - 1;
> >   	nbits = (size + (1UL << order) - 1) >> order;
> >   	rcu_read_lock();
> >   	list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { @@
> > -290,7 +294,7 @@ unsigned long gen_pool_alloc(struct gen_pool *pool,
> size_t size)
> >   		end_bit = chunk_size(chunk) >> order;
> >   retry:
> >   		start_bit = pool->algo(chunk->bits, end_bit, start_bit, nbits,
> > -				pool->data);
> > +				pool->data, align_mask);
> >   		if (start_bit >= end_bit)
> >   			continue;
> >   		remain = bitmap_set_ll(chunk->bits, start_bit, nbits); @@ -
> 309,6
> > +313,22 @@ retry:
> >   	rcu_read_unlock();
> >   	return addr;
> >   }
> > +EXPORT_SYMBOL(gen_pool_alloc_align);
> > +
> > +/**
> > + * gen_pool_alloc - allocate special memory from the pool
> > + * @pool: pool to allocate from
> > + * @size: number of bytes to allocate from the pool
> > + *
> > + * Allocate the requested number of bytes from the specified pool.
> > + * Uses the pool allocation function (with first-fit algorithm by
> default).
> > + * Can not be used in NMI handler on architectures without
> > + * NMI-safe cmpxchg implementation.
> > + */
> > +unsigned long gen_pool_alloc(struct gen_pool *pool, size_t size) {
> > +	return gen_pool_alloc_align(pool, size, 1);
> 
> Passing 1 here would change the behavior of the existing algorithms which
> were passing 0 for the align mask

When passing 1 here(align_mask = ((align + (1UL << order) - 1) >> order) - 1), align_mask will be 0.
It will not change the behavior of the existing algorithms.

> 
> > +}
> >   EXPORT_SYMBOL(gen_pool_alloc);
> >
> >   /**
> > @@ -502,9 +522,10 @@ EXPORT_SYMBOL(gen_pool_set_algo);
> >    * @data: additional data - unused
> >    */
> >   unsigned long gen_pool_first_fit(unsigned long *map, unsigned long
> size,
> > -		unsigned long start, unsigned int nr, void *data)
> > +		unsigned long start, unsigned int nr, void *data,
> > +		unsigned long align_mask)
> >   {
> > -	return bitmap_find_next_zero_area(map, size, start, nr, 0);
> > +	return bitmap_find_next_zero_area(map, size, start, nr, align_mask);
> >   }
> >   EXPORT_SYMBOL(gen_pool_first_fit);
> >
> > @@ -520,7 +541,7 @@ EXPORT_SYMBOL(gen_pool_first_fit);
> >    */
> >   unsigned long gen_pool_first_fit_order_align(unsigned long *map,
> >   		unsigned long size, unsigned long start,
> > -		unsigned int nr, void *data)
> > +		unsigned int nr, void *data, unsigned long align_mask)
> >   {
> >   	unsigned long align_mask = roundup_pow_of_two(nr) - 1;
> >
> > @@ -541,13 +562,14 @@ EXPORT_SYMBOL(gen_pool_first_fit_order_align);
> >    * which we can allocate the memory.
> >    */
> >   unsigned long gen_pool_best_fit(unsigned long *map, unsigned long
> size,
> > -		unsigned long start, unsigned int nr, void *data)
> > +				unsigned long start, unsigned int nr,
> > +				void *data, unsigned long align_mask)
> >   {
> >   	unsigned long start_bit = size;
> >   	unsigned long len = size + 1;
> >   	unsigned long index;
> >
> > -	index = bitmap_find_next_zero_area(map, size, start, nr, 0);
> > +	index = bitmap_find_next_zero_area(map, size, start, nr,
> > +align_mask);
> >
> >   	while (index < size) {
> >   		int next_bit = find_next_bit(map, size, index + nr);
> >
> 
> Thanks,
> Laura
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ