lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <871poz4983.fsf@mailhost.krisman.be>
Date: Mon, 25 Aug 2025 13:11:40 -0400
From: Gabriel Krisman Bertazi <gabriel@...sman.be>
To: Amir Goldstein <amir73il@...il.com>
Cc: André Almeida <andrealmeid@...lia.com>,  Miklos Szeredi
 <miklos@...redi.hu>,  Theodore Tso <tytso@....edu>,
  linux-unionfs@...r.kernel.org,  linux-kernel@...r.kernel.org,
  linux-fsdevel@...r.kernel.org,  Alexander Viro <viro@...iv.linux.org.uk>,
  Christian Brauner <brauner@...nel.org>,  Jan Kara <jack@...e.cz>,
  kernel-dev@...lia.com
Subject: Re: [PATCH v6 4/9] ovl: Create ovl_casefold() to support casefolded
 strncmp()

Amir Goldstein <amir73il@...il.com> writes:

> On Mon, Aug 25, 2025 at 5:27 PM Amir Goldstein <amir73il@...il.com> wrote:
>>
>> On Mon, Aug 25, 2025 at 1:09 PM Gabriel Krisman Bertazi
>> <gabriel@...sman.be> wrote:
>> >
>> > André Almeida <andrealmeid@...lia.com> writes:
>> >
>> > > To add overlayfs support casefold layers, create a new function
>> > > ovl_casefold(), to be able to do case-insensitive strncmp().
>> > >
>> > > ovl_casefold() allocates a new buffer and stores the casefolded version
>> > > of the string on it. If the allocation or the casefold operation fails,
>> > > fallback to use the original string.
>> > >
>> > > The case-insentive name is then used in the rb-tree search/insertion
>> > > operation. If the name is found in the rb-tree, the name can be
>> > > discarded and the buffer is freed. If the name isn't found, it's then
>> > > stored at struct ovl_cache_entry to be used later.
>> > >
>> > > Reviewed-by: Amir Goldstein <amir73il@...il.com>
>> > > Signed-off-by: André Almeida <andrealmeid@...lia.com>
>> > > ---
>> > > Changes from v6:
>> > >  - Last version was using `strncmp(... tmp->len)` which was causing
>> > >    regressions. It should be `strncmp(... len)`.
>> > >  - Rename cf_len to c_len
>> > >  - Use c_len for tree operation: (cmp < 0 || len < tmp->c_len)
>> > >  - Remove needless kfree(cf_name)
>> > > ---
>> > >  fs/overlayfs/readdir.c | 113 ++++++++++++++++++++++++++++++++++++++++---------
>> > >  1 file changed, 94 insertions(+), 19 deletions(-)
>> > >
>> > > diff --git a/fs/overlayfs/readdir.c b/fs/overlayfs/readdir.c
>> > > index b65cdfce31ce27172d28d879559f1008b9c87320..dfc661b7bc3f87efbf14991e97cee169400d823b 100644
>> > > --- a/fs/overlayfs/readdir.c
>> > > +++ b/fs/overlayfs/readdir.c
>> > > @@ -27,6 +27,8 @@ struct ovl_cache_entry {
>> > >       bool is_upper;
>> > >       bool is_whiteout;
>> > >       bool check_xwhiteout;
>> > > +     const char *c_name;
>> > > +     int c_len;
>> > >       char name[];
>> > >  };
>> > >
>> > > @@ -45,6 +47,7 @@ struct ovl_readdir_data {
>> > >       struct list_head *list;
>> > >       struct list_head middle;
>> > >       struct ovl_cache_entry *first_maybe_whiteout;
>> > > +     struct unicode_map *map;
>> > >       int count;
>> > >       int err;
>> > >       bool is_upper;
>> > > @@ -66,6 +69,27 @@ static struct ovl_cache_entry *ovl_cache_entry_from_node(struct rb_node *n)
>> > >       return rb_entry(n, struct ovl_cache_entry, node);
>> > >  }
>> > >
>> > > +static int ovl_casefold(struct unicode_map *map, const char *str, int len, char **dst)
>> > > +{
>> > > +     const struct qstr qstr = { .name = str, .len = len };
>> > > +     int cf_len;
>> > > +
>> > > +     if (!IS_ENABLED(CONFIG_UNICODE) || !map || is_dot_dotdot(str, len))
>> > > +             return 0;
>> > > +
>> > > +     *dst = kmalloc(NAME_MAX, GFP_KERNEL);
>> > > +
>> > > +     if (dst) {
>> > > +             cf_len = utf8_casefold(map, &qstr, *dst, NAME_MAX);
>> > > +
>> > > +             if (cf_len > 0)
>> > > +                     return cf_len;
>> > > +     }
>> > > +
>> > > +     kfree(*dst);
>> > > +     return 0;
>> > > +}
>> >
>> > Hi,
>> >
>> > I should just note this does not differentiates allocation errors from
>> > casefolding errors (invalid encoding).  It might be just a theoretical
>> > error because GFP_KERNEL shouldn't fail (wink, wink) and the rest of the
>> > operation is likely to fail too, but if you have an allocation failure, you
>> > can end up with an inconsistent cache, because a file is added under the
>> > !casefolded name and a later successful lookup will look for the
>> > casefolded version.
>>
>> Good point.
>> I will fix this in my tree.
>
> wait why should we not fail to fill the cache for both allocation
> and encoding errors?
>

We shouldn't fail the cache for encoding errors, just for allocation errors.

Perhaps I am misreading the code, so please correct me if I'm wrong.  if
ovl_casefold fails, the non-casefolded name is used in the cache.  That
makes sense if the reason utf8_casefold failed is because the string
cannot be casefolded (i.e. an invalid utf-8 string). For those strings,
everything is fine.  But on an allocation failure, the string might have
a real casefolded version.  If we fallback to the original string as the
key, a cache lookup won't find the entry, since we compare with memcmp.

>
>>
>> >
>> > > +
>> > >  static bool ovl_cache_entry_find_link(const char *name, int len,
>> > >                                     struct rb_node ***link,
>> > >                                     struct rb_node **parent)
>> > > @@ -79,10 +103,10 @@ static bool ovl_cache_entry_find_link(const char *name, int len,
>> > >
>> > >               *parent = *newp;
>> > >               tmp = ovl_cache_entry_from_node(*newp);
>> > > -             cmp = strncmp(name, tmp->name, len);
>> > > +             cmp = strncmp(name, tmp->c_name, len);
>> > >               if (cmp > 0)
>> > >                       newp = &tmp->node.rb_right;
>> > > -             else if (cmp < 0 || len < tmp->len)
>> > > +             else if (cmp < 0 || len < tmp->c_len)
>> > >                       newp = &tmp->node.rb_left;
>> > >               else
>> > >                       found = true;
>> > > @@ -101,10 +125,10 @@ static struct ovl_cache_entry *ovl_cache_entry_find(struct rb_root *root,
>> > >       while (node) {
>> > >               struct ovl_cache_entry *p = ovl_cache_entry_from_node(node);
>> > >
>> > > -             cmp = strncmp(name, p->name, len);
>> > > +             cmp = strncmp(name, p->c_name, len);
>> > >               if (cmp > 0)
>> > >                       node = p->node.rb_right;
>> > > -             else if (cmp < 0 || len < p->len)
>> > > +             else if (cmp < 0 || len < p->c_len)
>> > >                       node = p->node.rb_left;
>> > >               else
>> > >                       return p;
>> > > @@ -145,6 +169,7 @@ static bool ovl_calc_d_ino(struct ovl_readdir_data *rdd,
>> > >
>> > >  static struct ovl_cache_entry *ovl_cache_entry_new(struct ovl_readdir_data *rdd,
>> > >                                                  const char *name, int len,
>> > > +                                                const char *c_name, int c_len,
>> > >                                                  u64 ino, unsigned int d_type)
>> > >  {
>> > >       struct ovl_cache_entry *p;
>> > > @@ -167,6 +192,14 @@ static struct ovl_cache_entry *ovl_cache_entry_new(struct ovl_readdir_data *rdd,
>> > >       /* Defer check for overlay.whiteout to ovl_iterate() */
>> > >       p->check_xwhiteout = rdd->in_xwhiteouts_dir && d_type == DT_REG;
>> > >
>> > > +     if (c_name && c_name != name) {
>> > > +             p->c_name = c_name;
>> > > +             p->c_len = c_len;
>> > > +     } else {
>> > > +             p->c_name = p->name;
>> > > +             p->c_len = len;
>> > > +     }
>> > > +
>> > >       if (d_type == DT_CHR) {
>> > >               p->next_maybe_whiteout = rdd->first_maybe_whiteout;
>> > >               rdd->first_maybe_whiteout = p;
>> > > @@ -174,48 +207,55 @@ static struct ovl_cache_entry *ovl_cache_entry_new(struct ovl_readdir_data *rdd,
>> > >       return p;
>> > >  }
>> > >
>> > > -static bool ovl_cache_entry_add_rb(struct ovl_readdir_data *rdd,
>> > > -                               const char *name, int len, u64 ino,
>> > > +/* Return 0 for found, 1 for added, <0 for error */
>> > > +static int ovl_cache_entry_add_rb(struct ovl_readdir_data *rdd,
>> > > +                               const char *name, int len,
>> > > +                               const char *c_name, int c_len,
>> > > +                               u64 ino,
>> > >                                 unsigned int d_type)
>> > >  {
>> > >       struct rb_node **newp = &rdd->root->rb_node;
>> > >       struct rb_node *parent = NULL;
>> > >       struct ovl_cache_entry *p;
>> > >
>> > > -     if (ovl_cache_entry_find_link(name, len, &newp, &parent))
>> > > -             return true;
>> > > +     if (ovl_cache_entry_find_link(c_name, c_len, &newp, &parent))
>> > > +             return 0;
>> > >
>> > > -     p = ovl_cache_entry_new(rdd, name, len, ino, d_type);
>> > > +     p = ovl_cache_entry_new(rdd, name, len, c_name, c_len, ino, d_type);
>> > >       if (p == NULL) {
>> > >               rdd->err = -ENOMEM;
>> > > -             return false;
>> > > +             return -ENOMEM;
>> > >       }
>> > >
>> > >       list_add_tail(&p->l_node, rdd->list);
>> > >       rb_link_node(&p->node, parent, newp);
>> > >       rb_insert_color(&p->node, rdd->root);
>> > >
>> > > -     return true;
>> > > +     return 1;
>> > >  }
>> > >
>> > > -static bool ovl_fill_lowest(struct ovl_readdir_data *rdd,
>> > > +/* Return 0 for found, 1 for added, <0 for error */
>> > > +static int ovl_fill_lowest(struct ovl_readdir_data *rdd,
>> > >                          const char *name, int namelen,
>> > > +                        const char *c_name, int c_len,
>> > >                          loff_t offset, u64 ino, unsigned int d_type)
>> > >  {
>> > >       struct ovl_cache_entry *p;
>> > >
>> > > -     p = ovl_cache_entry_find(rdd->root, name, namelen);
>> > > +     p = ovl_cache_entry_find(rdd->root, c_name, c_len);
>> > >       if (p) {
>> > >               list_move_tail(&p->l_node, &rdd->middle);
>> > > +             return 0;
>> > >       } else {
>> > > -             p = ovl_cache_entry_new(rdd, name, namelen, ino, d_type);
>> > > +             p = ovl_cache_entry_new(rdd, name, namelen, c_name, c_len,
>> > > +                                     ino, d_type);
>> > >               if (p == NULL)
>> > >                       rdd->err = -ENOMEM;
>> > >               else
>> > >                       list_add_tail(&p->l_node, &rdd->middle);
>> > >       }
>> > >
>> > > -     return rdd->err == 0;
>> > > +     return rdd->err ?: 1;
>> > >  }
>> > >
>> > >  void ovl_cache_free(struct list_head *list)
>> > > @@ -223,8 +263,11 @@ void ovl_cache_free(struct list_head *list)
>> > >       struct ovl_cache_entry *p;
>> > >       struct ovl_cache_entry *n;
>> > >
>> > > -     list_for_each_entry_safe(p, n, list, l_node)
>> > > +     list_for_each_entry_safe(p, n, list, l_node) {
>> > > +             if (p->c_name != p->name)
>> > > +                     kfree(p->c_name);
>> > >               kfree(p);
>> > > +     }
>> > >
>> > >       INIT_LIST_HEAD(list);
>> > >  }
>> > > @@ -260,12 +303,36 @@ static bool ovl_fill_merge(struct dir_context *ctx, const char *name,
>> > >  {
>> > >       struct ovl_readdir_data *rdd =
>> > >               container_of(ctx, struct ovl_readdir_data, ctx);
>> > > +     struct ovl_fs *ofs = OVL_FS(rdd->dentry->d_sb);
>> > > +     const char *c_name = NULL;
>> > > +     char *cf_name = NULL;
>> > > +     int c_len = 0, ret;
>> > > +
>> > > +     if (ofs->casefold)
>> > > +             c_len = ovl_casefold(rdd->map, name, namelen, &cf_name);
>> > > +
>> > > +     if (c_len <= 0) {
>> > > +             c_name = name;
>> > > +             c_len = namelen;
>> > > +     } else {
>> > > +             c_name = cf_name;
>> > > +     }
>> > >
>> > >       rdd->count++;
>> > >       if (!rdd->is_lowest)
>> > > -             return ovl_cache_entry_add_rb(rdd, name, namelen, ino, d_type);
>> > > +             ret = ovl_cache_entry_add_rb(rdd, name, namelen, c_name, c_len, ino, d_type);
>> > >       else
>> > > -             return ovl_fill_lowest(rdd, name, namelen, offset, ino, d_type);
>> > > +             ret = ovl_fill_lowest(rdd, name, namelen, c_name, c_len, offset, ino, d_type);
>> > > +
>> > > +     /*
>> > > +      * If ret == 1, that means that c_name is being used as part of struct
>> > > +      * ovl_cache_entry and will be freed at ovl_cache_free(). Otherwise,
>> > > +      * c_name was found in the rb-tree so we can free it here.
>> > > +      */
>> > > +     if (ret != 1 && c_name != name)
>> > > +             kfree(c_name);
>> > > +
>> >
>> > The semantics of this being conditionally freed is a bit annoying, as
>> > it is already replicated in 3 places. I suppose a helper would come in
>> > hand.
>>
>> Yeh.
>>
>> I have already used ovl_cache_entry_free() in my tree.
>>
>> Thanks,
>> Amir.
>>
>> >
>> > In this specific case, it could just be:
>> >
>> > if (ret != 1)
>> >         kfree(cf_name);
>> >
>> >
>> > > +     return ret >= 0;
>> > >  }
>> > >
>> > >  static int ovl_check_whiteouts(const struct path *path, struct ovl_readdir_data *rdd)
>> > > @@ -357,12 +424,18 @@ static int ovl_dir_read_merged(struct dentry *dentry, struct list_head *list,
>> > >               .list = list,
>> > >               .root = root,
>> > >               .is_lowest = false,
>> > > +             .map = NULL,
>> > >       };
>> > >       int idx, next;
>> > >       const struct ovl_layer *layer;
>> > > +     struct ovl_fs *ofs = OVL_FS(dentry->d_sb);
>> > >
>> > >       for (idx = 0; idx != -1; idx = next) {
>> > >               next = ovl_path_next(idx, dentry, &realpath, &layer);
>> > > +
>> > > +             if (ofs->casefold)
>> > > +                     rdd.map = sb_encoding(realpath.dentry->d_sb);
>> > > +
>> > >               rdd.is_upper = ovl_dentry_upper(dentry) == realpath.dentry;
>> > >               rdd.in_xwhiteouts_dir = layer->has_xwhiteouts &&
>> > >                                       ovl_dentry_has_xwhiteouts(dentry);
>> > > @@ -555,7 +628,7 @@ static bool ovl_fill_plain(struct dir_context *ctx, const char *name,
>> > >               container_of(ctx, struct ovl_readdir_data, ctx);
>> > >
>> > >       rdd->count++;
>> > > -     p = ovl_cache_entry_new(rdd, name, namelen, ino, d_type);
>> > > +     p = ovl_cache_entry_new(rdd, name, namelen, NULL, 0, ino, d_type);
>> > >       if (p == NULL) {
>> > >               rdd->err = -ENOMEM;
>> > >               return false;
>> > > @@ -1023,6 +1096,8 @@ int ovl_check_empty_dir(struct dentry *dentry, struct list_head *list)
>> > >
>> > >  del_entry:
>> > >               list_del(&p->l_node);
>> > > +             if (p->c_name != p->name)
>> > > +                     kfree(p->c_name);
>> > >               kfree(p);
>> > >       }
>> >
>> > --
>> > Gabriel Krisman Bertazi

-- 
Gabriel Krisman Bertazi

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ