[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1b748d8d-44b0-6b79-9fa7-7ab6ee681f4b@themaw.net>
Date: Fri, 15 Sep 2023 09:05:09 +0800
From: Ian Kent <raven@...maw.net>
To: Amir Goldstein <amir73il@...il.com>,
Miklos Szeredi <mszeredi@...hat.com>
Cc: linux-fsdevel@...r.kernel.org, linux-kernel@...r.kernel.org,
linux-api@...r.kernel.org, linux-man@...r.kernel.org,
linux-security-module@...r.kernel.org, Karel Zak <kzak@...hat.com>,
David Howells <dhowells@...hat.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Al Viro <viro@...iv.linux.org.uk>,
Christian Brauner <christian@...uner.io>
Subject: Re: [RFC PATCH 2/3] add statmnt(2) syscall
On 14/9/23 14:11, Amir Goldstein wrote:
> On Wed, Sep 13, 2023 at 6:22 PM Miklos Szeredi <mszeredi@...hat.com> wrote:
>> Add a way to query attributes of a single mount instead of having to parse
>> the complete /proc/$PID/mountinfo, which might be huge.
>>
>> Lookup the mount by the old (32bit) or new (64bit) mount ID. If a mount
>> needs to be queried based on path, then statx(2) can be used to first query
>> the mount ID belonging to the path.
>>
>> Design is based on a suggestion by Linus:
>>
>> "So I'd suggest something that is very much like "statfsat()", which gets
>> a buffer and a length, and returns an extended "struct statfs" *AND*
>> just a string description at the end."
>>
>> The interface closely mimics that of statx.
>>
>> Handle ASCII attributes by appending after the end of the structure (as per
>> above suggestion). Allow querying multiple string attributes with
>> individual offset/length for each. String are nul terminated (termination
>> isn't counted in length).
>>
>> Mount options are also delimited with nul characters. Unlike proc, special
>> characters are not quoted.
>>
>> Link: https://lore.kernel.org/all/CAHk-=wh5YifP7hzKSbwJj94+DZ2czjrZsczy6GBimiogZws=rg@mail.gmail.com/
>> Signed-off-by: Miklos Szeredi <mszeredi@...hat.com>
>> ---
>> arch/x86/entry/syscalls/syscall_64.tbl | 1 +
>> fs/internal.h | 5 +
>> fs/namespace.c | 312 ++++++++++++++++++++++++-
>> fs/proc_namespace.c | 19 +-
>> fs/statfs.c | 1 +
>> include/linux/syscalls.h | 3 +
>> include/uapi/asm-generic/unistd.h | 5 +-
>> include/uapi/linux/mount.h | 36 +++
>> 8 files changed, 373 insertions(+), 9 deletions(-)
>>
>> diff --git a/arch/x86/entry/syscalls/syscall_64.tbl b/arch/x86/entry/syscalls/syscall_64.tbl
>> index 1d6eee30eceb..6d807c30cd16 100644
>> --- a/arch/x86/entry/syscalls/syscall_64.tbl
>> +++ b/arch/x86/entry/syscalls/syscall_64.tbl
>> @@ -375,6 +375,7 @@
>> 451 common cachestat sys_cachestat
>> 452 common fchmodat2 sys_fchmodat2
>> 453 64 map_shadow_stack sys_map_shadow_stack
>> +454 common statmnt sys_statmnt
>>
>> #
>> # Due to a historical design error, certain syscalls are numbered differently
>> diff --git a/fs/internal.h b/fs/internal.h
>> index d64ae03998cc..8f75271428aa 100644
>> --- a/fs/internal.h
>> +++ b/fs/internal.h
>> @@ -83,6 +83,11 @@ int path_mount(const char *dev_name, struct path *path,
>> const char *type_page, unsigned long flags, void *data_page);
>> int path_umount(struct path *path, int flags);
>>
>> +/*
>> + * proc_namespace.c
>> + */
>> +int show_path(struct seq_file *m, struct dentry *root);
>> +
>> /*
>> * fs_struct.c
>> */
>> diff --git a/fs/namespace.c b/fs/namespace.c
>> index de47c5f66e17..088a52043bba 100644
>> --- a/fs/namespace.c
>> +++ b/fs/namespace.c
>> @@ -69,7 +69,8 @@ static DEFINE_IDA(mnt_id_ida);
>> static DEFINE_IDA(mnt_group_ida);
>>
>> /* Don't allow confusion with mount ID allocated wit IDA */
>> -static atomic64_t mnt_id_ctr = ATOMIC64_INIT(1ULL << 32);
>> +#define OLD_MNT_ID_MAX UINT_MAX
>> +static atomic64_t mnt_id_ctr = ATOMIC64_INIT(OLD_MNT_ID_MAX);
>>
>> static struct hlist_head *mount_hashtable __read_mostly;
>> static struct hlist_head *mountpoint_hashtable __read_mostly;
>> @@ -4678,6 +4679,315 @@ SYSCALL_DEFINE5(mount_setattr, int, dfd, const char __user *, path,
>> return err;
>> }
>>
>> +static bool mnt_id_match(struct mount *mnt, u64 id)
>> +{
>> + if (id <= OLD_MNT_ID_MAX)
>> + return id == mnt->mnt_id;
>> + else
>> + return id == mnt->mnt_id_unique;
>> +}
>> +
>> +struct vfsmount *lookup_mnt_in_ns(u64 id, struct mnt_namespace *ns)
>> +{
>> + struct mount *mnt;
>> + struct vfsmount *res = NULL;
>> +
>> + lock_ns_list(ns);
>> + list_for_each_entry(mnt, &ns->list, mnt_list) {
>> + if (!mnt_is_cursor(mnt) && mnt_id_match(mnt, id)) {
>> + res = &mnt->mnt;
>> + break;
>> + }
>> + }
>> + unlock_ns_list(ns);
>> + return res;
>> +}
>> +
>> +struct stmt_state {
>> + void __user *const buf;
>> + size_t const bufsize;
>> + struct vfsmount *const mnt;
>> + u64 const mask;
>> + struct seq_file seq;
>> + struct path root;
>> + struct statmnt sm;
>> + size_t pos;
>> + int err;
>> +};
>> +
>> +typedef int (*stmt_func_t)(struct stmt_state *);
>> +
>> +static int stmt_string_seq(struct stmt_state *s, stmt_func_t func)
>> +{
>> + struct seq_file *seq = &s->seq;
>> + int ret;
>> +
>> + seq->count = 0;
>> + seq->size = min_t(size_t, seq->size, s->bufsize - s->pos);
>> + seq->buf = kvmalloc(seq->size, GFP_KERNEL_ACCOUNT);
>> + if (!seq->buf)
>> + return -ENOMEM;
>> +
>> + ret = func(s);
>> + if (ret)
>> + return ret;
>> +
>> + if (seq_has_overflowed(seq)) {
>> + if (seq->size == s->bufsize - s->pos)
>> + return -EOVERFLOW;
>> + seq->size *= 2;
>> + if (seq->size > MAX_RW_COUNT)
>> + return -ENOMEM;
>> + kvfree(seq->buf);
>> + return 0;
>> + }
>> +
>> + /* Done */
>> + return 1;
>> +}
>> +
>> +static void stmt_string(struct stmt_state *s, u64 mask, stmt_func_t func,
>> + stmt_str_t *str)
>> +{
>> + int ret = s->pos >= s->bufsize ? -EOVERFLOW : 0;
>> + struct statmnt *sm = &s->sm;
>> + struct seq_file *seq = &s->seq;
>> +
>> + if (s->err || !(s->mask & mask))
>> + return;
>> +
>> + seq->size = PAGE_SIZE;
>> + while (!ret)
>> + ret = stmt_string_seq(s, func);
>> +
>> + if (ret < 0) {
>> + s->err = ret;
>> + } else {
>> + seq->buf[seq->count++] = '\0';
>> + if (copy_to_user(s->buf + s->pos, seq->buf, seq->count)) {
>> + s->err = -EFAULT;
>> + } else {
>> + str->off = s->pos;
>> + str->len = seq->count - 1;
>> + s->pos += seq->count;
>> + }
>> + }
>> + kvfree(seq->buf);
>> + sm->mask |= mask;
>> +}
>> +
>> +static void stmt_numeric(struct stmt_state *s, u64 mask, stmt_func_t func)
>> +{
>> + if (s->err || !(s->mask & mask))
>> + return;
>> +
>> + s->err = func(s);
>> + s->sm.mask |= mask;
>> +}
>> +
>> +static u64 mnt_to_attr_flags(struct vfsmount *mnt)
>> +{
>> + unsigned int mnt_flags = READ_ONCE(mnt->mnt_flags);
>> + u64 attr_flags = 0;
>> +
>> + if (mnt_flags & MNT_READONLY)
>> + attr_flags |= MOUNT_ATTR_RDONLY;
>> + if (mnt_flags & MNT_NOSUID)
>> + attr_flags |= MOUNT_ATTR_NOSUID;
>> + if (mnt_flags & MNT_NODEV)
>> + attr_flags |= MOUNT_ATTR_NODEV;
>> + if (mnt_flags & MNT_NOEXEC)
>> + attr_flags |= MOUNT_ATTR_NOEXEC;
>> + if (mnt_flags & MNT_NODIRATIME)
>> + attr_flags |= MOUNT_ATTR_NODIRATIME;
>> + if (mnt_flags & MNT_NOSYMFOLLOW)
>> + attr_flags |= MOUNT_ATTR_NOSYMFOLLOW;
>> +
>> + if (mnt_flags & MNT_NOATIME)
>> + attr_flags |= MOUNT_ATTR_NOATIME;
>> + else if (mnt_flags & MNT_RELATIME)
>> + attr_flags |= MOUNT_ATTR_RELATIME;
>> + else
>> + attr_flags |= MOUNT_ATTR_STRICTATIME;
>> +
>> + if (is_idmapped_mnt(mnt))
>> + attr_flags |= MOUNT_ATTR_IDMAP;
>> +
>> + return attr_flags;
>> +}
>> +
>> +static u64 mnt_to_propagation_flags(struct mount *m)
>> +{
>> + u64 propagation = 0;
>> +
>> + if (IS_MNT_SHARED(m))
>> + propagation |= MS_SHARED;
>> + if (IS_MNT_SLAVE(m))
>> + propagation |= MS_SLAVE;
>> + if (IS_MNT_UNBINDABLE(m))
>> + propagation |= MS_UNBINDABLE;
>> + if (!propagation)
>> + propagation |= MS_PRIVATE;
>> +
>> + return propagation;
>> +}
>> +
>> +static int stmt_sb_basic(struct stmt_state *s)
>> +{
>> + struct super_block *sb = s->mnt->mnt_sb;
>> +
>> + s->sm.sb_dev_major = MAJOR(sb->s_dev);
>> + s->sm.sb_dev_minor = MINOR(sb->s_dev);
>> + s->sm.sb_magic = sb->s_magic;
>> + s->sm.sb_flags = sb->s_flags & (SB_RDONLY|SB_SYNCHRONOUS|SB_DIRSYNC|SB_LAZYTIME);
>> +
>> + return 0;
>> +}
>> +
>> +static int stmt_mnt_basic(struct stmt_state *s)
>> +{
>> + struct mount *m = real_mount(s->mnt);
>> +
>> + s->sm.mnt_id = m->mnt_id_unique;
>> + s->sm.mnt_parent_id = m->mnt_parent->mnt_id_unique;
>> + s->sm.mnt_id_old = m->mnt_id;
>> + s->sm.mnt_parent_id_old = m->mnt_parent->mnt_id;
>> + s->sm.mnt_attr = mnt_to_attr_flags(&m->mnt);
>> + s->sm.mnt_propagation = mnt_to_propagation_flags(m);
>> + s->sm.mnt_peer_group = IS_MNT_SHARED(m) ? m->mnt_group_id : 0;
>> + s->sm.mnt_master = IS_MNT_SLAVE(m) ? m->mnt_master->mnt_group_id : 0;
>> +
>> + return 0;
>> +}
>> +
>> +static int stmt_propagate_from(struct stmt_state *s)
>> +{
>> + struct mount *m = real_mount(s->mnt);
>> +
>> + if (!IS_MNT_SLAVE(m))
>> + return 0;
>> +
>> + s->sm.propagate_from = get_dominating_id(m, ¤t->fs->root);
>> +
>> + return 0;
>> +}
>> +
>> +static int stmt_mnt_root(struct stmt_state *s)
>> +{
>> + struct seq_file *seq = &s->seq;
>> + int err = show_path(seq, s->mnt->mnt_root);
>> +
>> + if (!err && !seq_has_overflowed(seq)) {
>> + seq->buf[seq->count] = '\0';
>> + seq->count = string_unescape_inplace(seq->buf, UNESCAPE_OCTAL);
>> + }
>> + return err;
>> +}
>> +
>> +static int stmt_mountpoint(struct stmt_state *s)
>> +{
>> + struct vfsmount *mnt = s->mnt;
>> + struct path mnt_path = { .dentry = mnt->mnt_root, .mnt = mnt };
>> + int err = seq_path_root(&s->seq, &mnt_path, &s->root, "");
>> +
>> + return err == SEQ_SKIP ? 0 : err;
>> +}
>> +
>> +static int stmt_fs_type(struct stmt_state *s)
>> +{
>> + struct seq_file *seq = &s->seq;
>> + struct super_block *sb = s->mnt->mnt_sb;
>> +
>> + seq_puts(seq, sb->s_type->name);
>> + if (sb->s_subtype) {
>> + seq_putc(seq, '.');
>> + seq_puts(seq, sb->s_subtype);
>> + }
>> + return 0;
>> +}
>> +
>> +static int stmt_sb_opts(struct stmt_state *s)
>> +{
>> + struct seq_file *seq = &s->seq;
>> + struct super_block *sb = s->mnt->mnt_sb;
>> + char *p, *end, *next, *u = seq->buf;
>> + int err;
>> +
>> + if (!sb->s_op->show_options)
>> + return 0;
>> +
>> + err = sb->s_op->show_options(seq, s->mnt->mnt_root);
>> + if (err || seq_has_overflowed(seq) || !seq->count)
>> + return err;
>> +
>> + end = seq->buf + seq->count;
>> + *end = '\0';
>> + for (p = seq->buf + 1; p < end; p = next + 1) {
>> + next = strchrnul(p, ',');
>> + *next = '\0';
>> + u += string_unescape(p, u, 0, UNESCAPE_OCTAL) + 1;
>> + }
>> + seq->count = u - 1 - seq->buf;
>> + return 0;
>> +}
>> +
>> +static int do_statmnt(struct stmt_state *s)
>> +{
>> + struct statmnt *sm = &s->sm;
>> + struct mount *m = real_mount(s->mnt);
>> +
>> + if (!capable(CAP_SYS_ADMIN) &&
>> + !is_path_reachable(m, m->mnt.mnt_root, &s->root))
>> + return -EPERM;
>> +
>> + stmt_numeric(s, STMT_SB_BASIC, stmt_sb_basic);
>> + stmt_numeric(s, STMT_MNT_BASIC, stmt_mnt_basic);
>> + stmt_numeric(s, STMT_PROPAGATE_FROM, stmt_propagate_from);
>> + stmt_string(s, STMT_MNT_ROOT, stmt_mnt_root, &sm->mnt_root);
>> + stmt_string(s, STMT_MOUNTPOINT, stmt_mountpoint, &sm->mountpoint);
>> + stmt_string(s, STMT_FS_TYPE, stmt_fs_type, &sm->fs_type);
>> + stmt_string(s, STMT_SB_OPTS, stmt_sb_opts, &sm->sb_opts);
>> +
>> + if (s->err)
>> + return s->err;
>> +
>> + if (copy_to_user(s->buf, sm, min_t(size_t, s->bufsize, sizeof(*sm))))
>> + return -EFAULT;
>> +
>> + return 0;
> Similar concern as with listmnt, I think that users would
> want to have a way to get the fixed size statmnt part that fits
> in the buffer, even if the variable length string values do not fit
> and be able to query the required buffer size to get the strings.
>
> The API could be either to explicitly request
> STMT_MNT_ROOT_LEN | STMT_MOUNTPOINT_LEN ...
> without allowing mixing of no-value and value requests,
> or to out-out from any string values using a single flag,
> which is probably more simple for API and implementation.
There is also the possibility that the size needed to satisfy the
request will change between request and call, not sure how to deal
with that, but the size estimate is needed ...
Ian
Powered by blists - more mailing lists