lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20061204200645.GN4392@kernel.dk>
Date:	Mon, 4 Dec 2006 21:06:46 +0100
From:	Jens Axboe <jens.axboe@...cle.com>
To:	"Chen, Kenneth W" <kenneth.w.chen@...el.com>
Cc:	linux-kernel <linux-kernel@...r.kernel.org>
Subject: Re: [patch] speed up single bio_vec allocation

On Mon, Dec 04 2006, Chen, Kenneth W wrote:
> On 64-bit arch like x86_64, struct bio is 104 byte.  Since bio slab is
> created with SLAB_HWCACHE_ALIGN flag, there are usually spare memory
> available at the end of bio.  I think we can utilize that memory for
> bio_vec allocation.  The purpose is not so much on saving memory consumption
> for bio_vec, instead, I'm attempting to optimize away a call to bvec_alloc_bs.
> 
> So here is a patch to do just that for 1 segment bio_vec (we currently only
> have space for 1 on 2.6.19).  And the detection whether there are spare space
> available is dynamically calculated at compile time.  If there are no space
> available, there will be no run time cost at all because gcc simply optimize
> away all the code added in this patch.  If there are space available, the only
> run time check is to see what the size of iovec is and we do appropriate
> assignment to bio->bi_io_Vec etc.  The cost is minimal and we gain a whole
> lot back from not calling bvec_alloc_bs() function.
> 
> I tried to use cache_line_size() to find out the alignment of struct bio, but
> stumbled on that it is a runtime function for x86_64. So instead I made bio
> to hint to the slab allocator to align on 32 byte (slab will use the larger
> value of hw cache line and caller hints of "align").  I think it is a sane
> number for majority of the CPUs out in the world.

Any benchmarks for this one?

> --- ./fs/bio.c.orig	2006-12-03 17:20:36.000000000 -0800
> +++ ./fs/bio.c	2006-12-03 21:29:20.000000000 -0800
> @@ -29,11 +29,14 @@
>  #include <scsi/sg.h>		/* for struct sg_iovec */
>  
>  #define BIO_POOL_SIZE 256
> -
> +#define BIO_ALIGN 32		/* minimal bio structure alignment */
>  static kmem_cache_t *bio_slab __read_mostly;
>  
>  #define BIOVEC_NR_POOLS 6
>  
> +#define BIOVEC_FIT_INSIDE_BIO_CACHE_LINE				\
> +	(ALIGN(sizeof(struct bio), BIO_ALIGN) ==			\
> +	 ALIGN(sizeof(struct bio) + sizeof(struct bio_vec), BIO_ALIGN))
>  /*
>   * a small number of entries is fine, not going to be performance critical.
>   * basically we just need to survive
> @@ -113,7 +116,8 @@ void bio_free(struct bio *bio, struct bi
>  
>  	BIO_BUG_ON(pool_idx >= BIOVEC_NR_POOLS);
>  
> -	mempool_free(bio->bi_io_vec, bio_set->bvec_pools[pool_idx]);
> +	if (!BIOVEC_FIT_INSIDE_BIO_CACHE_LINE || pool_idx)
> +		mempool_free(bio->bi_io_vec, bio_set->bvec_pools[pool_idx]);

I think you should use a flag for this instead.

>  	mempool_free(bio, bio_set->bio_pool);
>  }
>  
> @@ -166,7 +170,15 @@ struct bio *bio_alloc_bioset(gfp_t gfp_m
>  		struct bio_vec *bvl = NULL;
>  
>  		bio_init(bio);
> -		if (likely(nr_iovecs)) {
> +
> +		/*
> +		 * if bio_vec can fit into remaining cache line of struct
> +		 * bio, go ahead use it and skip mempool allocation.
> +		 */
> +		if (nr_iovecs == 1 && BIOVEC_FIT_INSIDE_BIO_CACHE_LINE) {
> +			bvl = (struct bio_vec*) (bio + 1);
> +			bio->bi_max_vecs = 1;
> +		} else if (likely(nr_iovecs)) {
>  			unsigned long idx = 0; /* shut up gcc */

Either move this logic to bvec_alloc_bs(), or remember to clear bvl as
is done there.

> @@ -1204,7 +1216,7 @@ static void __init biovec_init_slabs(voi
>  		struct biovec_slab *bvs = bvec_slabs + i;
>  
>  		size = bvs->nr_vecs * sizeof(struct bio_vec);
> -		bvs->slab = kmem_cache_create(bvs->name, size, 0,
> +		bvs->slab = kmem_cache_create(bvs->name, size, BIO_ALIGN,
>                                  SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL, NULL);
>  	}
>  }

Another idea would be to kill SLAB_HWCACHE_ALIGN (it's pretty pointless,
I bet), and always alloc sizeof(*bio) + sizeof(*bvl) in one go when a
bio is allocated. It doesn't add a lot of overhead even for the case
where we do > 1 page bios, and it gets rid of the dual allocation for
the 1 page bio.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ