lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 10 Jan 2011 09:30:41 -0700
From:	Andreas Dilger <adilger@...ger.ca>
To:	Lukas Czerner <lczerner@...hat.com>
Cc:	"linux-ext4@...r.kernel.org" <linux-ext4@...r.kernel.org>,
	"tytso@....edu" <tytso@....edu>,
	"lczerner@...hat.com" <lczerner@...hat.com>,
	"sandeen@...hat.com" <sandeen@...hat.com>,
	"kalpak@...geny.com" <kalpak@...geny.com>
Subject: Re: [PATCH 2/2] e2fsprogs: Add rbtree backend for bitmaps, use it instead of bitarrays

On 2011-01-10, at 8:18, Lukas Czerner <lczerner@...hat.com> wrote:
> This commit adds another backend to store bitmaps. It is based on
> rbtrees and it stores just used extents of bitmaps. It means that it can
> be more memory efficient in most cases and also with a careful use it
> might be much faster than simple bitarrays.
> 
> 
> @@ -66,7 +66,7 @@ errcode_t ext2fs_allocate_inode_bitmap(ext2_filsys fs,
>    if (fs->flags & EXT2_FLAG_64BITS)
>        return (ext2fs_alloc_generic_bmap(fs,
>                  EXT2_ET_MAGIC_INODE_BITMAP64,
> -                  EXT2FS_BMAP64_BITARRAY,
> +                  EXT2FS_BMAP64_RBTREE,
>                  start, end, real_end, descr, ret));

It would be really useful to allow runtime selection of b
>    /* Otherwise, check to see if the file system is small enough
> @@ -98,7 +98,7 @@ errcode_t ext2fs_allocate_block_bitmap(ext2_filsys fs,
>    if (fs->flags & EXT2_FLAG_64BITS)
>        return (ext2fs_alloc_generic_bmap(fs,
>                  EXT2_ET_MAGIC_BLOCK_BITMAP64,
> -                  EXT2FS_BMAP64_BITARRAY,
> +                  EXT2FS_BMAP64_RBTREE,
>                  start, end, real_end, descr, ret));
> 
>    if ((end > ~0U) || (real_end > ~0U))
> diff --git a/lib/ext2fs/blkmap64_rb.c b/lib/ext2fs/blkmap64_rb.c
> new file mode 100644
> index 0000000..6de8eb9
> --- /dev/null
> +++ b/lib/ext2fs/blkmap64_rb.c
> @@ -0,0 +1,815 @@
> +/*
> + * blkmap64_rb.c --- Simple rb-tree implementation for bitmaps
> + *
> + * (C)2010 Red Hat, Inc., Lukas Czerner <lczerner@...hat.com>
> + *
> + * %Begin-Header%
> + * This file may be redistributed under the terms of the GNU Public
> + * License.
> + * %End-Header%
> + */
> +
> +#include <stdio.h>
> +#include <string.h>
> +#if HAVE_UNISTD_H
> +#include <unistd.h>
> +#endif
> +#include <fcntl.h>
> +#include <time.h>
> +#if HAVE_SYS_STAT_H
> +#include <sys/stat.h>
> +#endif
> +#if HAVE_SYS_TYPES_H
> +#include <sys/types.h>
> +#endif
> +
> +#include "ext2_fs.h"
> +#include "ext2fsP.h"
> +#include "bmap64.h"
> +#include "rbtree.h"
> +
> +#include <limits.h>
> +
> +struct bmap_rb_extent {
> +    struct rb_node node;
> +    __u64 start;
> +    __u32 count;
> +};
> +
> +struct ext2fs_rb_private {
> +    struct rb_root root;
> +    struct bmap_rb_extent *cache;
> +};
> +
> +static int _rb_insert_extent(struct bmap_rb_extent *, struct rb_root *, int);
> +static int rb_insert_extent(struct bmap_rb_extent *,
> +                struct ext2fs_rb_private *);
> +static void rb_flush_cache(struct ext2fs_rb_private *);
> +static int rb_get_new_extent(struct bmap_rb_extent **, __u64, __u64);
> +
> +/*#define DEBUG*/
> +
> +#ifdef DEBUG
> +static void print_tree(struct rb_root *root)
> +{
> +    struct rb_node *node = NULL;
> +    struct bmap_rb_extent *ext;
> +
> +    printf("\t\t\t=================================\n");
> +    node = rb_first(root);
> +    for (node = rb_first(root); node != NULL; node=rb_next(node)) {
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +        printf("\t\t\t--> (%llu -> %llu)\n",
> +            ext->start, ext->start + ext->count);
> +    }
> +    printf("\t\t\t=================================\n");
> +}
> +
> +static int check_tree(struct rb_root *root, const char *msg)
> +{
> +    struct rb_node *new_node, *node, *next;
> +    struct bmap_rb_extent *ext, *old = NULL;
> +
> +    for (node = rb_first(root); node; node = rb_next(node)) {
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +        if (ext->count <= 0) {
> +            printf("Tree Error: count is crazy\n");
> +            printf("extent: %llu -> %llu (%llu)\n", ext->start,
> +                ext->start + ext->count, ext->count);
> +            goto err_out;
> +        }
> +        if (ext->start < 0) {
> +            printf("Tree Error: start is crazy\n");
> +            printf("extent: %llu -> %llu (%llu)\n", ext->start,
> +                ext->start + ext->count, ext->count);
> +            goto err_out;
> +        }
> +
> +        if (old) {
> +            if (old->start > ext->start) {
> +                printf("Tree Error: start is crazy\n");
> +                printf("extent: %llu -> %llu (%llu)\n",
> +                    old->start, old->start + old->count,
> +                    old->count);
> +                printf("extent next: %llu -> %llu (%llu)\n",
> +                    ext->start, ext->start + ext->count,
> +                    ext->count);
> +                goto err_out;
> +            }
> +            if ((old->start + old->count) >= ext->start) {
> +                printf("Tree Error: extent is crazy\n");
> +                printf("extent: %llu -> %llu (%llu)\n",
> +                    old->start, old->start + old->count,
> +                    old->count);
> +                printf("extent next: %llu -> %llu (%llu)\n",
> +                    ext->start, ext->start + ext->count,
> +                    ext->count);
> +                goto err_out;
> +            }
> +        }
> +        old = ext;
> +    }
> +    return 0;
> +
> +err_out:
> +    printf("%s\n", msg);
> +    print_tree(root);
> +    exit(1);
> +}
> +#else
> +#define check_tree(root, msg) 0
> +#define print_tree(root, msg) 0
> +#endif
> +
> +static int rb_get_new_extent(struct bmap_rb_extent **ext, __u64 start,
> +                 __u64 count)
> +{
> +    struct bmap_rb_extent *new_ext;
> +    int retval = 0;
> +
> +    retval = ext2fs_get_mem(sizeof (struct bmap_rb_extent),
> +                &new_ext);
> +    /*
> +     * Not quite sure how to do this, but passing the error up the stack
> +     * probably is not the best idea.
> +     */
> +    if (retval) {
> +        fprintf(stderr, "Error: NOT ENOUGH MEMORY!\n");
> +        exit(ENOMEM);
> +    }
> +
> +    new_ext->start = start;
> +    new_ext->count = count;
> +    *ext = new_ext;
> +    return retval;
> +}
> +
> +static void rb_flush_cache(struct ext2fs_rb_private *bp)
> +{
> +    if (!bp->cache)
> +        return;
> +
> +    _rb_insert_extent(bp->cache, &bp->root, 0);
> +    check_tree(&bp->root, __func__);
> +    bp->cache = NULL;
> +}
> +
> +/*
> + * Wrapper around _rb_insert_extent.
> + * First check cache, when it is emtpy put ext into cache, in the other
> + * case, try to merge ext with cache. If they are mutually exclusive
> + * insert old cache into tree and put ext into cache.
> + */
> +static int rb_insert_extent(struct bmap_rb_extent *ext,
> +                struct ext2fs_rb_private *bp)
> +{
> +    struct bmap_rb_extent *cache = bp->cache;
> +    __u64 cend, eend;
> +    int ret = 1;
> +
> +    if (!cache) {
> +        bp->cache = ext;
> +        return 0;
> +    }
> +
> +    cend = cache->start + cache->count;
> +    eend = ext->start + ext->count;
> +
> +    /* Mutually exclusive extents */
> +    if ((ext->start > cend) ||
> +        (cache->start > (ext->start + ext->count))) {
> +        ret = _rb_insert_extent(bp->cache, &bp->root, 0);
> +        bp->cache = ext;
> +        check_tree(&bp->root, __func__);
> +        return ret;
> +    }
> +
> +    if (ext->start < cache->start) {
> +        /* embedded interval */
> +        if (cend >= eend)
> +            return 1;
> +        cache->count += cache->start - ext->start;
> +        cache->start = ext->start;
> +    } else
> +        cache->count += (eend - cend);
> +
> +
> +    bp->cache = cache;
> +    return ret;
> +}
> +
> +static errcode_t rb_alloc_private_data (ext2fs_generic_bitmap bitmap)
> +{
> +    struct ext2fs_rb_private *bp;
> +    errcode_t    retval;
> +
> +    retval = ext2fs_get_mem(sizeof (struct ext2fs_rb_private), &bp);
> +    if (retval)
> +        return retval;
> +
> +    bp->root = RB_ROOT;
> +    bp->cache = NULL;
> +
> +
> +    bitmap->private = (void *) bp;
> +    return 0;
> +}
> +
> +static errcode_t rb_new_bmap(ext2_filsys fs EXT2FS_ATTR((unused)),
> +                 ext2fs_generic_bitmap bitmap)
> +{
> +    errcode_t    retval;
> +
> +    retval = rb_alloc_private_data (bitmap);
> +    if (retval)
> +        return retval;
> +
> +    return 0;
> +}
> +
> +static void rb_free_tree(struct rb_root *root)
> +{
> +    struct bmap_rb_extent *ext;
> +    struct rb_node *node;
> +    int count = 0;
> +
> +    node = rb_first(root);
> +    while (node) {
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +        rb_erase(node, root);
> +        ext2fs_free_mem(&ext);
> +        node = rb_first(root);
> +        count++;
> +    }
> +}
> +
> +static void rb_free_bmap(ext2fs_generic_bitmap bitmap)
> +{
> +    struct ext2fs_rb_private *bp;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +
> +    rb_free_tree(&bp->root);
> +    ext2fs_free_mem(&bp->cache);
> +    ext2fs_free_mem(&bp);
> +    bp = 0;
> +}
> +
> +static errcode_t rb_copy_bmap(ext2fs_generic_bitmap src,
> +                  ext2fs_generic_bitmap dest)
> +{
> +    struct ext2fs_rb_private *src_bp, *dest_bp;
> +    struct bmap_rb_extent *src_ext, *dest_ext;
> +    struct rb_node *dest_node, *src_node, *dest_last, **n;
> +    errcode_t retval = 0;
> +
> +    retval = rb_alloc_private_data (dest);
> +    if (retval)
> +        return retval;
> +
> +    src_bp = (struct ext2fs_rb_private *) src->private;
> +    dest_bp = (struct ext2fs_rb_private *) dest->private;
> +    rb_flush_cache(src_bp);
> +
> +    src_node = rb_first(&src_bp->root);
> +    while (src_node) {
> +        src_ext = rb_entry(src_node, struct bmap_rb_extent, node);
> +        retval = ext2fs_get_mem(sizeof (struct bmap_rb_extent),
> +                    &dest_ext);
> +        if (retval)
> +            break;
> +
> +        memcpy(dest_ext, src_ext, sizeof(struct bmap_rb_extent));
> +
> +        dest_node = &dest_ext->node;
> +        n = &dest_bp->root.rb_node;
> +
> +        dest_last = NULL;
> +        if (*n) {
> +            dest_last = rb_last(&dest_bp->root);
> +            n = &(dest_last)->rb_right;
> +        }
> +
> +        rb_link_node(dest_node, dest_last, n);
> +        rb_insert_color(dest_node, &dest_bp->root);
> +
> +        src_node = rb_next(src_node);
> +    }
> +
> +    return retval;
> +}
> +
> +static void rb_truncate(__u64 new_max, struct rb_root *root)
> +{
> +    struct bmap_rb_extent *ext;
> +    struct rb_node *node;
> +
> +    node = rb_last(root);
> +    while (node) {
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +
> +        if ((ext->start + ext->count - 1) <= new_max)
> +            break;
> +        else if (ext->start > new_max) {
> +            rb_erase(node, root);
> +            ext2fs_free_mem(&ext);
> +            node = rb_last(root);
> +            continue;
> +        } else
> +            ext->count = new_max - ext->start + 1;
> +    }
> +}
> +
> +static errcode_t rb_resize_bmap(ext2fs_generic_bitmap bmap,
> +                __u64 new_end, __u64 new_real_end)
> +{
> +    struct ext2fs_rb_private *bp;
> +
> +    if (new_real_end >= bmap->real_end) {
> +        bmap->end = new_end;
> +        bmap->real_end = new_real_end;
> +        return 0;
> +    }
> +
> +    bp = (struct ext2fs_rb_private *) bmap->private;
> +    rb_flush_cache(bp);
> +
> +    /* truncate tree to new_real_end size */
> +    rb_truncate(new_real_end, &bp->root);
> +
> +    bmap->end = new_end;
> +    bmap->real_end = new_real_end;
> +    return 0;
> +
> +}
> +
> +/*
> + * It would be nice to have read cache for this, so when walking bitmap
> + * in sequential manner we do not have to go through the list for every bit.
> + *
> + * The idea is:
> + * 1. check if there is a cache.
> + * 2. look for match in the cached extent
> + * 3. if not found, search the tree.
> + * 4. put found extent into the cache.
> + */
> +static int
> +rb_test_bit(struct ext2fs_rb_private *bp, __u64 bit)
> +{
> +    struct rb_node *parent = NULL;
> +    struct rb_node **n = &bp->root.rb_node;
> +    struct bmap_rb_extent *ext;
> +
> +    while (*n) {
> +        parent = *n;
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +        if (bit < ext->start)
> +            n = &(*n)->rb_left;
> +        else if (bit >= (ext->start + ext->count))
> +            n = &(*n)->rb_right;
> +        else
> +            return 1;
> +    }
> +    return 0;
> +}
> +
> +static int
> +rb_set_bit(struct ext2fs_rb_private *bp, __u64 bit)
> +{
> +    struct bmap_rb_extent *cache = bp->cache;
> +    struct bmap_rb_extent *ext;
> +    int retval;
> +
> +    if (!cache)
> +        goto new_cache;
> +
> +    if (bit == (cache->start + cache->count)) {
> +        cache->count++;
> +        return 0;
> +    }
> +
> +    /* Mutually exclusive */
> +    if ((bit > (cache->start + cache->count)) ||
> +        (bit < cache->start)) {
> +        _rb_insert_extent(bp->cache, &bp->root, 0);
> +        goto new_cache;
> +    }
> +
> +    return 1;
> +
> +new_cache:
> +    retval = rb_get_new_extent(&ext, bit, 1);
> +    if (retval)
> +        return retval;
> +    bp->cache = ext;
> +    return 0;
> +}
> +
> +static int _rb_insert_extent(struct bmap_rb_extent *new_ext,
> +                  struct rb_root *root, int in)
> +{
> +    struct rb_node *parent = NULL, **n = &root->rb_node;
> +    struct rb_node *new_node, *node, *next;
> +    struct bmap_rb_extent *ext;
> +    __u64 start, count, old_start, old_count;
> +    int retval = 0;
> +
> +    start = old_start = new_ext->start;
> +    count = old_count = new_ext->count;
> +
> +    while (*n) {
> +        parent = *n;
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +
> +        if (start < ext->start) {
> +            n = &(*n)->rb_left;
> +        } else if (start > (ext->start + ext->count)) {
> +            n = &(*n)->rb_right;
> +        } else {
> +            ext2fs_free_mem(&new_ext);
> +            if ((start + count) <= (ext->start + ext->count))
> +                return 1;
> +
> +            count += (start - ext->start);
> +            start = ext->start;
> +            new_ext = ext;
> +            new_node = &ext->node;
> +            retval = 1;
> +
> +            goto skip_insert;
> +        }
> +    }
> +
> +    new_node = &new_ext->node;
> +    rb_link_node(new_node, parent, n);
> +    rb_insert_color(new_node, root);
> +
> +    node = rb_prev(new_node);
> +    if (node) {
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +        if ((ext->start + ext->count) == start) {
> +            start = ext->start;
> +            count += ext->count;
> +            rb_erase(node, root);
> +            ext2fs_free_mem(&ext);
> +        }
> +    }
> +
> +skip_insert:
> +    /* See if we can merge extent to the right */
> +    for (node = rb_next(new_node); node != NULL; node = next) {
> +        next = rb_next(node);
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +
> +        if ((ext->start + ext->count) <= start)
> +            continue;
> +
> +        /* No more merging */
> +        if ((start + count) < ext->start)
> +            break;
> +
> +        /* ext is embedded in new_ext interval */
> +        if ((start + count) >= (ext->start + ext->count)) {
> +            rb_erase(node, root);
> +            ext2fs_free_mem(&ext);
> +            continue;
> +        } else {
> +        /* merge ext with new_ext */
> +            count += ((ext->start + ext->count) -
> +                  (start + count));
> +            rb_erase(node, root);
> +            ext2fs_free_mem(&ext);
> +            break;
> +        }
> +    }
> +
> +    new_ext->start = start;
> +    new_ext->count = count;
> +
> +    return retval;
> +}
> +
> +static int rb_remove_extent(struct bmap_rb_extent *del_ext,
> +                  struct rb_root *root)
> +{
> +    struct rb_node *parent = NULL, **n = &root->rb_node;
> +    struct rb_node *next;
> +    struct bmap_rb_extent *ext;
> +    __u64 start, count, old_count, old_start;
> +    int retval = 0;
> +
> +    if (RB_EMPTY_ROOT(root))
> +        return 0;
> +
> +    start = old_start = del_ext->start;
> +    count = old_count = del_ext->count;
> +
> +    while (*n) {
> +        parent = *n;
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +        if (start < ext->start) {
> +            n = &(*n)->rb_left;
> +            continue;
> +        } else if (start >= (ext->start + ext->count)) {
> +            n = &(*n)->rb_right;
> +            continue;
> +        }
> +
> +        if ((start > ext->start) &&
> +            (start + count) < (ext->start + ext->count)) {
> +            /* We have to split extent into two */
> +            del_ext->start = start + count;
> +            del_ext->count = (ext->start + ext->count) -
> +                     del_ext->start;
> +
> +            ext->count = start - ext->start;
> +            retval = _rb_insert_extent(del_ext, root, 0);
> +            if (retval)
> +                printf("\t%s Warning - extent should be "
> +                    "unique, but it is not\n", __func__);
> +            return 1;
> +        }
> +
> +        if ((start + count) >= (ext->start + ext->count))
> +            ext->count = start - ext->start;
> +
> +        if (0 == ext->count) {
> +            parent = rb_next(&ext->node);
> +            rb_erase(&ext->node, root);
> +            ext2fs_free_mem(&ext);
> +            goto search_right;
> +        }
> +
> +        if (start == ext->start) {
> +            ext->start += count;
> +            ext->count -= count;
> +            return retval;
> +        }
> +    }
> +
> +search_right:
> +    /* See if we should delete or truncate extent on the right */
> +    for (; parent != NULL; parent = next) {
> +        next = rb_next(parent);
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +        if ((ext->start + ext->count) <= start)
> +            continue;
> +
> +        /* No more merging */
> +        if ((start + count) < ext->start)
> +            break;
> +
> +        /* ext is embedded in new_ext interval */
> +        if ((start + count) >= (ext->start + ext->count)) {
> +            rb_erase(&ext->node, root);
> +            ext2fs_free_mem(&ext);
> +            continue;
> +        } else {
> +        /* merge ext with new_ext */
> +            ext->count -= ((start + count) - ext->start);
> +            ext->start = start + count;
> +            break;
> +        }
> +    }
> +
> +    return retval;
> +}
> +
> +static int rb_mark_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> +    struct ext2fs_rb_private *bp;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    arg -= bitmap->start;
> +
> +    return rb_set_bit(bp, arg);
> +}
> +
> +static int rb_unmark_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> +    struct ext2fs_rb_private *bp;
> +    struct bmap_rb_extent *del_ext;
> +    int retval;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    rb_flush_cache(bp);
> +    arg -= bitmap->start;
> +
> +    retval = rb_get_new_extent(&del_ext, arg, 1);
> +    if (retval)
> +        return retval;
> +
> +    retval = rb_remove_extent(del_ext, &bp->root);
> +    if (!retval)
> +        ext2fs_free_mem(&del_ext);
> +    check_tree(&bp->root, __func__);
> +
> +    return retval;
> +}
> +
> +static int rb_test_bmap(ext2fs_generic_bitmap bitmap, __u64 arg)
> +{
> +    struct ext2fs_rb_private *bp;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    rb_flush_cache(bp);
> +    arg -= bitmap->start;
> +
> +    return rb_test_bit(bp, arg);
> +}
> +
> +static void rb_mark_bmap_extent(ext2fs_generic_bitmap bitmap, __u64 arg,
> +                unsigned int num)
> +{
> +    struct ext2fs_rb_private *bp;
> +    struct bmap_rb_extent *new_ext;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    arg -= bitmap->start;
> +
> +    /* We should check and return exit value since allocation
> +     * may not be successful
> +     */
> +    rb_get_new_extent(&new_ext, arg, num);
> +    rb_insert_extent(new_ext, bp);
> +}
> +
> +static void rb_unmark_bmap_extent(ext2fs_generic_bitmap bitmap, __u64 arg,
> +                  unsigned int num)
> +{
> +    struct ext2fs_rb_private *bp;
> +    struct bmap_rb_extent *del_ext;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    rb_flush_cache(bp);
> +    arg -= bitmap->start;
> +
> +    /* We should check and return exit value since allocation
> +     * may not be successful
> +     */
> +    rb_get_new_extent(&del_ext, arg, num);
> +    rb_remove_extent(del_ext, &bp->root);
> +    check_tree(&bp->root, __func__);
> +}
> +
> +static int rb_test_clear_bmap_extent(ext2fs_generic_bitmap bitmap,
> +                     __u64 start, unsigned int len)
> +{
> +    struct rb_node *parent = NULL, **n;
> +    struct rb_node *node, *next;
> +    struct ext2fs_rb_private *bp;
> +    struct bmap_rb_extent *ext;
> +    int retval = 1;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    n = &bp->root.rb_node;
> +    rb_flush_cache(bp);
> +    start -= bitmap->start;
> +
> +    if ((len == 0) ||
> +        RB_EMPTY_ROOT(&bp->root))
> +        return 1;
> +
> +    /*
> +     * If we find nothing, we should examine whole extent, but
> +     * when we find match, the extent is not clean, thus be return
> +     * false.
> +     */
> +    while (*n) {
> +        parent = *n;
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +        if (start < ext->start) {
> +            n = &(*n)->rb_left;
> +        } else if (start >= (ext->start + ext->count)) {
> +            n = &(*n)->rb_right;
> +        } else {
> +            /*
> +             * We found extent int the tree -> extent is not
> +             * clean
> +             */
> +            return 0;
> +        }
> +    }
> +
> +    node = parent;
> +    while (node) {
> +        next = rb_next(node);
> +        ext = rb_entry(node, struct bmap_rb_extent, node);
> +        node = next;
> +
> +        if ((ext->start + ext->count) <= start)
> +            continue;
> +
> +        /* No more merging */
> +        if ((start + len) <= ext->start)
> +            break;
> +
> +        retval = 0;
> +        break;
> +    }
> +    return retval;
> +}
> +
> +static errcode_t rb_set_bmap_range(ext2fs_generic_bitmap bitmap,
> +                     __u64 start, size_t num, void *in)
> +{
> +    struct ext2fs_rb_private *bp;
> +    size_t i;
> +    int ret;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    rb_flush_cache(bp);
> +
> +    for (i = 0; i < num; i++) {
> +        ret = ext2fs_test_bit(i, in);
> +        if (ret)
> +            rb_set_bit(bp, start + i - bitmap->start);
> +    }
> +
> +    return 0;
> +}
> +
> +static errcode_t rb_get_bmap_range(ext2fs_generic_bitmap bitmap,
> +                     __u64 start, size_t num, void *out)
> +{
> +
> +    struct rb_node *parent = NULL, *next, **n;
> +    struct ext2fs_rb_private *bp;
> +    struct bmap_rb_extent *ext;
> +    __u64 pos;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +    n = &bp->root.rb_node;
> +    rb_flush_cache(bp);
> +    start -= bitmap->start;
> +
> +    if (RB_EMPTY_ROOT(&bp->root)) {
> +        return 0;
> +    }
> +
> +    while (*n) {
> +        parent = *n;
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +        if (start < ext->start) {
> +            n = &(*n)->rb_left;
> +        } else if (start >= (ext->start + ext->count)) {
> +            n = &(*n)->rb_right;
> +        } else
> +            break;
> +    }
> +
> +    pos = start;
> +    for (; parent != NULL; parent = next) {
> +        next = rb_next(parent);
> +        ext = rb_entry(parent, struct bmap_rb_extent, node);
> +
> +        while (((pos - start) < num) &&
> +            (pos < ext->start)) {
> +            ext2fs_fast_clear_bit64((pos - start), out);
> +            pos++;
> +        }
> +
> +        if ((pos - start) >= num)
> +            return 0;
> +
> +        while (((pos - start) < num) &&
> +            (pos < (ext->start + ext->count))) {
> +            ext2fs_fast_set_bit64((pos - start), out);
> +            pos++;
> +        }
> +    }
> +
> +    while ((pos - start) < num) {
> +        ext2fs_fast_clear_bit64((pos - start), out);
> +        pos++;
> +    }
> +
> +    return 0;
> +}
> +
> +static void rb_clear_bmap(ext2fs_generic_bitmap bitmap)
> +{
> +    struct ext2fs_rb_private *bp;
> +
> +    bp = (struct ext2fs_rb_private *) bitmap->private;
> +
> +    rb_free_tree(&bp->root);
> +    ext2fs_free_mem(&bp->cache);
> +}
> +
> +struct ext2_bitmap_ops ext2fs_blkmap64_rbtree = {
> +    .type = EXT2FS_BMAP64_RBTREE,
> +    .new_bmap = rb_new_bmap,
> +    .free_bmap = rb_free_bmap,
> +    .copy_bmap = rb_copy_bmap,
> +    .resize_bmap = rb_resize_bmap,
> +    .mark_bmap = rb_mark_bmap,
> +    .unmark_bmap = rb_unmark_bmap,
> +    .test_bmap = rb_test_bmap,
> +    .test_clear_bmap_extent = rb_test_clear_bmap_extent,
> +    .mark_bmap_extent = rb_mark_bmap_extent,
> +    .unmark_bmap_extent = rb_unmark_bmap_extent,
> +    .set_bmap_range = rb_set_bmap_range,
> +    .get_bmap_range = rb_get_bmap_range,
> +    .clear_bmap = rb_clear_bmap,
> +};
> diff --git a/lib/ext2fs/bmap64.h b/lib/ext2fs/bmap64.h
> index b0aa84c..31021b9 100644
> --- a/lib/ext2fs/bmap64.h
> +++ b/lib/ext2fs/bmap64.h
> @@ -59,3 +59,4 @@ struct ext2_bitmap_ops {
> };
> 
> extern struct ext2_bitmap_ops ext2fs_blkmap64_bitarray;
> +extern struct ext2_bitmap_ops ext2fs_blkmap64_rbtree;
> diff --git a/lib/ext2fs/ext2fsP.h b/lib/ext2fs/ext2fsP.h
> index ab9ee76..aa45c43 100644
> --- a/lib/ext2fs/ext2fsP.h
> +++ b/lib/ext2fs/ext2fsP.h
> @@ -108,6 +108,7 @@ extern void ext2fs_numeric_progress_close(ext2_filsys fs,
>  */
> 
> #define EXT2FS_BMAP64_BITARRAY    1
> +#define EXT2FS_BMAP64_RBTREE    2
> 
> extern errcode_t ext2fs_alloc_generic_bmap(ext2_filsys fs, errcode_t magic,
>                       int type, __u64 start, __u64 end,
> diff --git a/lib/ext2fs/gen_bitmap64.c b/lib/ext2fs/gen_bitmap64.c
> index df095ac..eb233f4 100644
> --- a/lib/ext2fs/gen_bitmap64.c
> +++ b/lib/ext2fs/gen_bitmap64.c
> @@ -91,6 +91,9 @@ errcode_t ext2fs_alloc_generic_bmap(ext2_filsys fs, errcode_t magic,
>    case EXT2FS_BMAP64_BITARRAY:
>        ops = &ext2fs_blkmap64_bitarray;
>        break;
> +    case EXT2FS_BMAP64_RBTREE:
> +        ops = &ext2fs_blkmap64_rbtree;
> +        break;
>    default:
>        return EINVAL;
>    }
> -- 
> 1.7.2.3
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ