[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <9268E343-A1FD-410D-8C4F-40976B233634@dilger.ca>
Date: Mon, 10 Jan 2011 09:30:41 -0700
From: Andreas Dilger <adilger@...ger.ca>
To: Lukas Czerner <lczerner@...hat.com>
Cc: "linux-ext4@...r.kernel.org" <linux-ext4@...r.kernel.org>,
"tytso@....edu" <tytso@....edu>,
"lczerner@...hat.com" <lczerner@...hat.com>,
"sandeen@...hat.com" <sandeen@...hat.com>,
"kalpak@...geny.com" <kalpak@...geny.com>
Subject: Re: [PATCH 2/2] e2fsprogs: Add rbtree backend for bitmaps, use it instead of bitarrays
On 2011-01-10, at 8:18, Lukas Czerner <lczerner@...hat.com> wrote:
> This commit adds another backend to store bitmaps. It is based on
> rbtrees and it stores just used extents of bitmaps. It means that it can
> be more memory efficient in most cases and also with a careful use it
> might be much faster than simple bitarrays.
>
>
> @@ -66,7 +66,7 @@ errcode_t ext2fs_allocate_inode_bitmap(ext2_filsys fs,
> if (fs->flags & EXT2_FLAG_64BITS)
> return (ext2fs_alloc_generic_bmap(fs,
> EXT2_ET_MAGIC_INODE_BITMAP64,
> - EXT2FS_BMAP64_BITARRAY,
> + EXT2FS_BMAP64_RBTREE,
> start, end, real_end, descr, ret));
It would be really useful to allow runtime selection of b
> /* Otherwise, check to see if the file system is small enough
> @@ -98,7 +98,7 @@ errcode_t ext2fs_allocate_block_bitmap(ext2_filsys fs,
> if (fs->flags & EXT2_FLAG_64BITS)
> return (ext2fs_alloc_generic_bmap(fs,
> EXT2_ET_MAGIC_BLOCK_BITMAP64,
> - EXT2FS_BMAP64_BITARRAY,
> + EXT2FS_BMAP64_RBTREE,
> start, end, real_end, descr, ret));
>
> if ((end > ~0U) || (real_end > ~0U))
> diff --git a/lib/ext2fs/blkmap64_rb.c b/lib/ext2fs/blkmap64_rb.c
> new file mode 100644
> index 0000000..6de8eb9
> --- /dev/null
> +++ b/lib/ext2fs/blkmap64_rb.c
> @@ -0,0 +1,815 @@
> +/*
> + * blkmap64_rb.c --- Simple rb-tree implementation for bitmaps
> + *
> + * (C)2010 Red Hat, Inc., Lukas Czerner <lczerner@...hat.com>
> + *
> + * %Begin-Header%
> + * This file may be redistributed under the terms of the GNU Public
> + * License.
> + * %End-Header%
> + */
> +
> +#include <stdio.h>
> +#include <string.h>
> +#if HAVE_UNISTD_H
> +#include <unistd.h>
> +#endif
> +#include <fcntl.h>
> +#include <time.h>
> +#if HAVE_SYS_STAT_H
> +#include <sys/stat.h>
> +#endif
> +#if HAVE_SYS_TYPES_H
> +#include <sys/types.h>
> +#endif
> +
> +#include "ext2_fs.h"
> +#include "ext2fsP.h"
> +#include "bmap64.h"
> +#include "rbtree.h"
> +
> +#include <limits.h>
> +
> +struct bmap_rb_extent {
> + struct rb_node node;
> + __u64 start;
> + __u32 count;
> +};
> +
> +struct ext2fs_rb_private {
> + struct rb_root root;
> + struct bmap_rb_extent *cache;
> +};
> +
> +static int _rb_insert_extent(struct bmap_rb_extent *, struct rb_root *, int);
> +static int rb_insert_extent(struct bmap_rb_extent *,
> + struct ext2fs_rb_private *);
> +static void rb_flush_cache(struct ext2fs_rb_private *);
> +static int rb_get_new_extent(struct bmap_rb_extent **, __u64, __u64);
> +
> +/*#define DEBUG*/
> +
> +#ifdef DEBUG
> +static void print_tree(struct rb_root *root)
> +{
> + struct rb_node *node = NULL;
> + struct bmap_rb_extent *ext;
> +
> + printf("\t\t\t=================================\n");
> + node = rb_first(root);
> + for (node = rb_first(root); node != NULL; node=rb_next(node)) {
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> + printf("\t\t\t--> (%llu -> %llu)\n",
> + ext->start, ext->start + ext->count);
> + }
> + printf("\t\t\t=================================\n");
> +}
> +
> +static int check_tree(struct rb_root *root, const char *msg)
> +{
> + struct rb_node *new_node, *node, *next;
> + struct bmap_rb_extent *ext, *old = NULL;
> +
> + for (node = rb_first(root); node; node = rb_next(node)) {
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> + if (ext->count <= 0) {
> + printf("Tree Error: count is crazy\n");
> + printf("extent: %llu -> %llu (%llu)\n", ext->start,
> + ext->start + ext->count, ext->count);
> + goto err_out;
> + }
> + if (ext->start < 0) {
> + printf("Tree Error: start is crazy\n");
> + printf("extent: %llu -> %llu (%llu)\n", ext->start,
> + ext->start + ext->count, ext->count);
> + goto err_out;
> + }
> +
> + if (old) {
> + if (old->start > ext->start) {
> + printf("Tree Error: start is crazy\n");
> + printf("extent: %llu -> %llu (%llu)\n",
> + old->start, old->start + old->count,
> + old->count);
> + printf("extent next: %llu -> %llu (%llu)\n",
> + ext->start, ext->start + ext->count,
> + ext->count);
> + goto err_out;
> + }
> + if ((old->start + old->count) >= ext->start) {
> + printf("Tree Error: extent is crazy\n");
> + printf("extent: %llu -> %llu (%llu)\n",
> + old->start, old->start + old->count,
> + old->count);
> + printf("extent next: %llu -> %llu (%llu)\n",
> + ext->start, ext->start + ext->count,
> + ext->count);
> + goto err_out;
> + }
> + }
> + old = ext;
> + }
> + return 0;
> +
> +err_out:
> + printf("%s\n", msg);
> + print_tree(root);
> + exit(1);
> +}
> +#else
> +#define check_tree(root, msg) 0
> +#define print_tree(root, msg) 0
> +#endif
> +
> +static int rb_get_new_extent(struct bmap_rb_extent **ext, __u64 start,
> + __u64 count)
> +{
> + struct bmap_rb_extent *new_ext;
> + int retval = 0;
> +
> + retval = ext2fs_get_mem(sizeof (struct bmap_rb_extent),
> + &new_ext);
> + /*
> + * Not quite sure how to do this, but passing the error up the stack
> + * probably is not the best idea.
> + */
> + if (retval) {
> + fprintf(stderr, "Error: NOT ENOUGH MEMORY!\n");
> + exit(ENOMEM);
> + }
> +
> + new_ext->start = start;
> + new_ext->count = count;
> + *ext = new_ext;
> + return retval;
> +}
> +
> +static void rb_flush_cache(struct ext2fs_rb_private *bp)
> +{
> + if (!bp->cache)
> + return;
> +
> + _rb_insert_extent(bp->cache, &bp->root, 0);
> + check_tree(&bp->root, __func__);
> + bp->cache = NULL;
> +}
> +
> +/*
> + * Wrapper around _rb_insert_extent.
> + * First check cache, when it is emtpy put ext into cache, in the other
> + * case, try to merge ext with cache. If they are mutually exclusive
> + * insert old cache into tree and put ext into cache.
> + */
> +static int rb_insert_extent(struct bmap_rb_extent *ext,
> + struct ext2fs_rb_private *bp)
> +{
> + struct bmap_rb_extent *cache = bp->cache;
> + __u64 cend, eend;
> + int ret = 1;
> +
> + if (!cache) {
> + bp->cache = ext;
> + return 0;
> + }
> +
> + cend = cache->start + cache->count;
> + eend = ext->start + ext->count;
> +
> + /* Mutually exclusive extents */
> + if ((ext->start > cend) ||
> + (cache->start > (ext->start + ext->count))) {
> + ret = _rb_insert_extent(bp->cache, &bp->root, 0);
> + bp->cache = ext;
> + check_tree(&bp->root, __func__);
> + return ret;
> + }
> +
> + if (ext->start < cache->start) {
> + /* embedded interval */
> + if (cend >= eend)
> + return 1;
> + cache->count += cache->start - ext->start;
> + cache->start = ext->start;
> + } else
> + cache->count += (eend - cend);
> +
> +
> + bp->cache = cache;
> + return ret;
> +}
> +
> +static errcode_t rb_alloc_private_data (ext2fs_generic_bitmap bitmap)
> +{
> + struct ext2fs_rb_private *bp;
> + errcode_t retval;
> +
> + retval = ext2fs_get_mem(sizeof (struct ext2fs_rb_private), &bp);
> + if (retval)
> + return retval;
> +
> + bp->root = RB_ROOT;
> + bp->cache = NULL;
> +
> +
> + bitmap->private = (void *) bp;
> + return 0;
> +}
> +
> +static errcode_t rb_new_bmap(ext2_filsys fs EXT2FS_ATTR((unused)),
> + ext2fs_generic_bitmap bitmap)
> +{
> + errcode_t retval;
> +
> + retval = rb_alloc_private_data (bitmap);
> + if (retval)
> + return retval;
> +
> + return 0;
> +}
> +
> +static void rb_free_tree(struct rb_root *root)
> +{
> + struct bmap_rb_extent *ext;
> + struct rb_node *node;
> + int count = 0;
> +
> + node = rb_first(root);
> + while (node) {
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> + rb_erase(node, root);
> + ext2fs_free_mem(&ext);
> + node = rb_first(root);
> + count++;
> + }
> +}
> +
> +static void rb_free_bmap(ext2fs_generic_bitmap bitmap)
> +{
> + struct ext2fs_rb_private *bp;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> +
> + rb_free_tree(&bp->root);
> + ext2fs_free_mem(&bp->cache);
> + ext2fs_free_mem(&bp);
> + bp = 0;
> +}
> +
> +static errcode_t rb_copy_bmap(ext2fs_generic_bitmap src,
> + ext2fs_generic_bitmap dest)
> +{
> + struct ext2fs_rb_private *src_bp, *dest_bp;
> + struct bmap_rb_extent *src_ext, *dest_ext;
> + struct rb_node *dest_node, *src_node, *dest_last, **n;
> + errcode_t retval = 0;
> +
> + retval = rb_alloc_private_data (dest);
> + if (retval)
> + return retval;
> +
> + src_bp = (struct ext2fs_rb_private *) src->private;
> + dest_bp = (struct ext2fs_rb_private *) dest->private;
> + rb_flush_cache(src_bp);
> +
> + src_node = rb_first(&src_bp->root);
> + while (src_node) {
> + src_ext = rb_entry(src_node, struct bmap_rb_extent, node);
> + retval = ext2fs_get_mem(sizeof (struct bmap_rb_extent),
> + &dest_ext);
> + if (retval)
> + break;
> +
> + memcpy(dest_ext, src_ext, sizeof(struct bmap_rb_extent));
> +
> + dest_node = &dest_ext->node;
> + n = &dest_bp->root.rb_node;
> +
> + dest_last = NULL;
> + if (*n) {
> + dest_last = rb_last(&dest_bp->root);
> + n = &(dest_last)->rb_right;
> + }
> +
> + rb_link_node(dest_node, dest_last, n);
> + rb_insert_color(dest_node, &dest_bp->root);
> +
> + src_node = rb_next(src_node);
> + }
> +
> + return retval;
> +}
> +
> +static void rb_truncate(__u64 new_max, struct rb_root *root)
> +{
> + struct bmap_rb_extent *ext;
> + struct rb_node *node;
> +
> + node = rb_last(root);
> + while (node) {
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> +
> + if ((ext->start + ext->count - 1) <= new_max)
> + break;
> + else if (ext->start > new_max) {
> + rb_erase(node, root);
> + ext2fs_free_mem(&ext);
> + node = rb_last(root);
> + continue;
> + } else
> + ext->count = new_max - ext->start + 1;
> + }
> +}
> +
> +static errcode_t rb_resize_bmap(ext2fs_generic_bitmap bmap,
> + __u64 new_end, __u64 new_real_end)
> +{
> + struct ext2fs_rb_private *bp;
> +
> + if (new_real_end >= bmap->real_end) {
> + bmap->end = new_end;
> + bmap->real_end = new_real_end;
> + return 0;
> + }
> +
> + bp = (struct ext2fs_rb_private *) bmap->private;
> + rb_flush_cache(bp);
> +
> + /* truncate tree to new_real_end size */
> + rb_truncate(new_real_end, &bp->root);
> +
> + bmap->end = new_end;
> + bmap->real_end = new_real_end;
> + return 0;
> +
> +}
> +
> +/*
> + * It would be nice to have read cache for this, so when walking bitmap
> + * in sequential manner we do not have to go through the list for every bit.
> + *
> + * The idea is:
> + * 1. check if there is a cache.
> + * 2. look for match in the cached extent
> + * 3. if not found, search the tree.
> + * 4. put found extent into the cache.
> + */
> +static int
> +rb_test_bit(struct ext2fs_rb_private *bp, __u64 bit)
> +{
> + struct rb_node *parent = NULL;
> + struct rb_node **n = &bp->root.rb_node;
> + struct bmap_rb_extent *ext;
> +
> + while (*n) {
> + parent = *n;
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> + if (bit < ext->start)
> + n = &(*n)->rb_left;
> + else if (bit >= (ext->start + ext->count))
> + n = &(*n)->rb_right;
> + else
> + return 1;
> + }
> + return 0;
> +}
> +
> +static int
> +rb_set_bit(struct ext2fs_rb_private *bp, __u64 bit)
> +{
> + struct bmap_rb_extent *cache = bp->cache;
> + struct bmap_rb_extent *ext;
> + int retval;
> +
> + if (!cache)
> + goto new_cache;
> +
> + if (bit == (cache->start + cache->count)) {
> + cache->count++;
> + return 0;
> + }
> +
> + /* Mutually exclusive */
> + if ((bit > (cache->start + cache->count)) ||
> + (bit < cache->start)) {
> + _rb_insert_extent(bp->cache, &bp->root, 0);
> + goto new_cache;
> + }
> +
> + return 1;
> +
> +new_cache:
> + retval = rb_get_new_extent(&ext, bit, 1);
> + if (retval)
> + return retval;
> + bp->cache = ext;
> + return 0;
> +}
> +
> +static int _rb_insert_extent(struct bmap_rb_extent *new_ext,
> + struct rb_root *root, int in)
> +{
> + struct rb_node *parent = NULL, **n = &root->rb_node;
> + struct rb_node *new_node, *node, *next;
> + struct bmap_rb_extent *ext;
> + __u64 start, count, old_start, old_count;
> + int retval = 0;
> +
> + start = old_start = new_ext->start;
> + count = old_count = new_ext->count;
> +
> + while (*n) {
> + parent = *n;
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> +
> + if (start < ext->start) {
> + n = &(*n)->rb_left;
> + } else if (start > (ext->start + ext->count)) {
> + n = &(*n)->rb_right;
> + } else {
> + ext2fs_free_mem(&new_ext);
> + if ((start + count) <= (ext->start + ext->count))
> + return 1;
> +
> + count += (start - ext->start);
> + start = ext->start;
> + new_ext = ext;
> + new_node = &ext->node;
> + retval = 1;
> +
> + goto skip_insert;
> + }
> + }
> +
> + new_node = &new_ext->node;
> + rb_link_node(new_node, parent, n);
> + rb_insert_color(new_node, root);
> +
> + node = rb_prev(new_node);
> + if (node) {
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> + if ((ext->start + ext->count) == start) {
> + start = ext->start;
> + count += ext->count;
> + rb_erase(node, root);
> + ext2fs_free_mem(&ext);
> + }
> + }
> +
> +skip_insert:
> + /* See if we can merge extent to the right */
> + for (node = rb_next(new_node); node != NULL; node = next) {
> + next = rb_next(node);
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> +
> + if ((ext->start + ext->count) <= start)
> + continue;
> +
> + /* No more merging */
> + if ((start + count) < ext->start)
> + break;
> +
> + /* ext is embedded in new_ext interval */
> + if ((start + count) >= (ext->start + ext->count)) {
> + rb_erase(node, root);
> + ext2fs_free_mem(&ext);
> + continue;
> + } else {
> + /* merge ext with new_ext */
> + count += ((ext->start + ext->count) -
> + (start + count));
> + rb_erase(node, root);
> + ext2fs_free_mem(&ext);
> + break;
> + }
> + }
> +
> + new_ext->start = start;
> + new_ext->count = count;
> +
> + return retval;
> +}
> +
> +static int rb_remove_extent(struct bmap_rb_extent *del_ext,
> + struct rb_root *root)
> +{
> + struct rb_node *parent = NULL, **n = &root->rb_node;
> + struct rb_node *next;
> + struct bmap_rb_extent *ext;
> + __u64 start, count, old_count, old_start;
> + int retval = 0;
> +
> + if (RB_EMPTY_ROOT(root))
> + return 0;
> +
> + start = old_start = del_ext->start;
> + count = old_count = del_ext->count;
> +
> + while (*n) {
> + parent = *n;
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> + if (start < ext->start) {
> + n = &(*n)->rb_left;
> + continue;
> + } else if (start >= (ext->start + ext->count)) {
> + n = &(*n)->rb_right;
> + continue;
> + }
> +
> + if ((start > ext->start) &&
> + (start + count) < (ext->start + ext->count)) {
> + /* We have to split extent into two */
> + del_ext->start = start + count;
> + del_ext->count = (ext->start + ext->count) -
> + del_ext->start;
> +
> + ext->count = start - ext->start;
> + retval = _rb_insert_extent(del_ext, root, 0);
> + if (retval)
> + printf("\t%s Warning - extent should be "
> + "unique, but it is not\n", __func__);
> + return 1;
> + }
> +
> + if ((start + count) >= (ext->start + ext->count))
> + ext->count = start - ext->start;
> +
> + if (0 == ext->count) {
> + parent = rb_next(&ext->node);
> + rb_erase(&ext->node, root);
> + ext2fs_free_mem(&ext);
> + goto search_right;
> + }
> +
> + if (start == ext->start) {
> + ext->start += count;
> + ext->count -= count;
> + return retval;
> + }
> + }
> +
> +search_right:
> + /* See if we should delete or truncate extent on the right */
> + for (; parent != NULL; parent = next) {
> + next = rb_next(parent);
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> + if ((ext->start + ext->count) <= start)
> + continue;
> +
> + /* No more merging */
> + if ((start + count) < ext->start)
> + break;
> +
> + /* ext is embedded in new_ext interval */
> + if ((start + count) >= (ext->start + ext->count)) {
> + rb_erase(&ext->node, root);
> + ext2fs_free_mem(&ext);
> + continue;
> + } else {
> + /* merge ext with new_ext */
> + ext->count -= ((start + count) - ext->start);
> + ext->start = start + count;
> + break;
> + }
> + }
> +
> + return retval;
> +}
> +
> +static int rb_mark_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> + struct ext2fs_rb_private *bp;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + arg -= bitmap->start;
> +
> + return rb_set_bit(bp, arg);
> +}
> +
> +static int rb_unmark_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> + struct ext2fs_rb_private *bp;
> + struct bmap_rb_extent *del_ext;
> + int retval;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + rb_flush_cache(bp);
> + arg -= bitmap->start;
> +
> + retval = rb_get_new_extent(&del_ext, arg, 1);
> + if (retval)
> + return retval;
> +
> + retval = rb_remove_extent(del_ext, &bp->root);
> + if (!retval)
> + ext2fs_free_mem(&del_ext);
> + check_tree(&bp->root, __func__);
> +
> + return retval;
> +}
> +
> +static int rb_test_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> + struct ext2fs_rb_private *bp;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + rb_flush_cache(bp);
> + arg -= bitmap->start;
> +
> + return rb_test_bit(bp, arg);
> +}
> +
> +static void rb_mark_bmap_extent(ext2fs_generic_bitmap bitmap, __u64 arg,
> + unsigned int num)
> +{
> + struct ext2fs_rb_private *bp;
> + struct bmap_rb_extent *new_ext;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + arg -= bitmap->start;
> +
> + /* We should check and return exit value since allocation
> + * may not be successful
> + */
> + rb_get_new_extent(&new_ext, arg, num);
> + rb_insert_extent(new_ext, bp);
> +}
> +
> +static void rb_unmark_bmap_extent(ext2fs_generic_bitmap bitmap, __u64 arg,
> + unsigned int num)
> +{
> + struct ext2fs_rb_private *bp;
> + struct bmap_rb_extent *del_ext;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + rb_flush_cache(bp);
> + arg -= bitmap->start;
> +
> + /* We should check and return exit value since allocation
> + * may not be successful
> + */
> + rb_get_new_extent(&del_ext, arg, num);
> + rb_remove_extent(del_ext, &bp->root);
> + check_tree(&bp->root, __func__);
> +}
> +
> +static int rb_test_clear_bmap_extent(ext2fs_generic_bitmap bitmap,
> + __u64 start, unsigned int len)
> +{
> + struct rb_node *parent = NULL, **n;
> + struct rb_node *node, *next;
> + struct ext2fs_rb_private *bp;
> + struct bmap_rb_extent *ext;
> + int retval = 1;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + n = &bp->root.rb_node;
> + rb_flush_cache(bp);
> + start -= bitmap->start;
> +
> + if ((len == 0) ||
> + RB_EMPTY_ROOT(&bp->root))
> + return 1;
> +
> + /*
> + * If we find nothing, we should examine whole extent, but
> + * when we find match, the extent is not clean, thus be return
> + * false.
> + */
> + while (*n) {
> + parent = *n;
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> + if (start < ext->start) {
> + n = &(*n)->rb_left;
> + } else if (start >= (ext->start + ext->count)) {
> + n = &(*n)->rb_right;
> + } else {
> + /*
> + * We found extent int the tree -> extent is not
> + * clean
> + */
> + return 0;
> + }
> + }
> +
> + node = parent;
> + while (node) {
> + next = rb_next(node);
> + ext = rb_entry(node, struct bmap_rb_extent, node);
> + node = next;
> +
> + if ((ext->start + ext->count) <= start)
> + continue;
> +
> + /* No more merging */
> + if ((start + len) <= ext->start)
> + break;
> +
> + retval = 0;
> + break;
> + }
> + return retval;
> +}
> +
> +static errcode_t rb_set_bmap_range(ext2fs_generic_bitmap bitmap,
> + __u64 start, size_t num, void *in)
> +{
> + struct ext2fs_rb_private *bp;
> + size_t i;
> + int ret;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + rb_flush_cache(bp);
> +
> + for (i = 0; i < num; i++) {
> + ret = ext2fs_test_bit(i, in);
> + if (ret)
> + rb_set_bit(bp, start + i - bitmap->start);
> + }
> +
> + return 0;
> +}
> +
> +static errcode_t rb_get_bmap_range(ext2fs_generic_bitmap bitmap,
> + __u64 start, size_t num, void *out)
> +{
> +
> + struct rb_node *parent = NULL, *next, **n;
> + struct ext2fs_rb_private *bp;
> + struct bmap_rb_extent *ext;
> + __u64 pos;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> + n = &bp->root.rb_node;
> + rb_flush_cache(bp);
> + start -= bitmap->start;
> +
> + if (RB_EMPTY_ROOT(&bp->root)) {
> + return 0;
> + }
> +
> + while (*n) {
> + parent = *n;
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> + if (start < ext->start) {
> + n = &(*n)->rb_left;
> + } else if (start >= (ext->start + ext->count)) {
> + n = &(*n)->rb_right;
> + } else
> + break;
> + }
> +
> + pos = start;
> + for (; parent != NULL; parent = next) {
> + next = rb_next(parent);
> + ext = rb_entry(parent, struct bmap_rb_extent, node);
> +
> + while (((pos - start) < num) &&
> + (pos < ext->start)) {
> + ext2fs_fast_clear_bit64((pos - start), out);
> + pos++;
> + }
> +
> + if ((pos - start) >= num)
> + return 0;
> +
> + while (((pos - start) < num) &&
> + (pos < (ext->start + ext->count))) {
> + ext2fs_fast_set_bit64((pos - start), out);
> + pos++;
> + }
> + }
> +
> + while ((pos - start) < num) {
> + ext2fs_fast_clear_bit64((pos - start), out);
> + pos++;
> + }
> +
> + return 0;
> +}
> +
> +static void rb_clear_bmap(ext2fs_generic_bitmap bitmap)
> +{
> + struct ext2fs_rb_private *bp;
> +
> + bp = (struct ext2fs_rb_private *) bitmap->private;
> +
> + rb_free_tree(&bp->root);
> + ext2fs_free_mem(&bp->cache);
> +}
> +
> +struct ext2_bitmap_ops ext2fs_blkmap64_rbtree = {
> + .type = EXT2FS_BMAP64_RBTREE,
> + .new_bmap = rb_new_bmap,
> + .free_bmap = rb_free_bmap,
> + .copy_bmap = rb_copy_bmap,
> + .resize_bmap = rb_resize_bmap,
> + .mark_bmap = rb_mark_bmap,
> + .unmark_bmap = rb_unmark_bmap,
> + .test_bmap = rb_test_bmap,
> + .test_clear_bmap_extent = rb_test_clear_bmap_extent,
> + .mark_bmap_extent = rb_mark_bmap_extent,
> + .unmark_bmap_extent = rb_unmark_bmap_extent,
> + .set_bmap_range = rb_set_bmap_range,
> + .get_bmap_range = rb_get_bmap_range,
> + .clear_bmap = rb_clear_bmap,
> +};
> diff --git a/lib/ext2fs/bmap64.h b/lib/ext2fs/bmap64.h
> index b0aa84c..31021b9 100644
> --- a/lib/ext2fs/bmap64.h
> +++ b/lib/ext2fs/bmap64.h
> @@ -59,3 +59,4 @@ struct ext2_bitmap_ops {
> };
>
> extern struct ext2_bitmap_ops ext2fs_blkmap64_bitarray;
> +extern struct ext2_bitmap_ops ext2fs_blkmap64_rbtree;
> diff --git a/lib/ext2fs/ext2fsP.h b/lib/ext2fs/ext2fsP.h
> index ab9ee76..aa45c43 100644
> --- a/lib/ext2fs/ext2fsP.h
> +++ b/lib/ext2fs/ext2fsP.h
> @@ -108,6 +108,7 @@ extern void ext2fs_numeric_progress_close(ext2_filsys fs,
> */
>
> #define EXT2FS_BMAP64_BITARRAY 1
> +#define EXT2FS_BMAP64_RBTREE 2
>
> extern errcode_t ext2fs_alloc_generic_bmap(ext2_filsys fs, errcode_t magic,
> int type, __u64 start, __u64 end,
> diff --git a/lib/ext2fs/gen_bitmap64.c b/lib/ext2fs/gen_bitmap64.c
> index df095ac..eb233f4 100644
> --- a/lib/ext2fs/gen_bitmap64.c
> +++ b/lib/ext2fs/gen_bitmap64.c
> @@ -91,6 +91,9 @@ errcode_t ext2fs_alloc_generic_bmap(ext2_filsys fs, errcode_t magic,
> case EXT2FS_BMAP64_BITARRAY:
> ops = &ext2fs_blkmap64_bitarray;
> break;
> + case EXT2FS_BMAP64_RBTREE:
> + ops = &ext2fs_blkmap64_rbtree;
> + break;
> default:
> return EINVAL;
> }
> --
> 1.7.2.3
>
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists