[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <cdc1962d-386a-4b33-9876-8552fc7593b9@lucifer.local>
Date: Wed, 13 Aug 2025 14:44:32 +0100
From: Lorenzo Stoakes <lorenzo.stoakes@...cle.com>
To: Suren Baghdasaryan <surenb@...gle.com>
Cc: akpm@...ux-foundation.org, Liam.Howlett@...cle.com, david@...hat.com,
vbabka@...e.cz, peterx@...hat.com, jannh@...gle.com,
hannes@...xchg.org, mhocko@...nel.org, paulmck@...nel.org,
shuah@...nel.org, adobriyan@...il.com, brauner@...nel.org,
josef@...icpanda.com, yebin10@...wei.com, linux@...ssschuh.net,
willy@...radead.org, osalvador@...e.de, andrii@...nel.org,
ryan.roberts@....com, christophe.leroy@...roup.eu,
tjmercier@...gle.com, kaleshsingh@...gle.com, aha310510@...il.com,
linux-kernel@...r.kernel.org, linux-fsdevel@...r.kernel.org,
linux-mm@...ck.org, linux-kselftest@...r.kernel.org,
SeongJae Park <sj@...nel.org>
Subject: Re: [PATCH v4 2/3] fs/proc/task_mmu: factor out proc_maps_private
fields used by PROCMAP_QUERY
On Fri, Aug 08, 2025 at 08:28:48AM -0700, Suren Baghdasaryan wrote:
> Refactor struct proc_maps_private so that the fields used by PROCMAP_QUERY
> ioctl are moved into a separate structure. In the next patch this allows
> ioctl to reuse some of the functions used for reading /proc/pid/maps
> without using file->private_data. This prevents concurrent modification
> of file->private_data members by ioctl and /proc/pid/maps readers.
>
> The change is pure code refactoring and has no functional changes.
>
> Signed-off-by: Suren Baghdasaryan <surenb@...gle.com>
> Reviewed-by: Vlastimil Babka <vbabka@...e.cz>
> Acked-by: SeongJae Park <sj@...nel.org>
LGTM, so:
Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@...cle.com>
> ---
> fs/proc/internal.h | 15 +++++---
> fs/proc/task_mmu.c | 87 +++++++++++++++++++++++---------------------
> fs/proc/task_nommu.c | 14 +++----
> 3 files changed, 63 insertions(+), 53 deletions(-)
>
> diff --git a/fs/proc/internal.h b/fs/proc/internal.h
> index e737401d7383..d1598576506c 100644
> --- a/fs/proc/internal.h
> +++ b/fs/proc/internal.h
> @@ -378,16 +378,21 @@ extern void proc_self_init(void);
> * task_[no]mmu.c
> */
> struct mem_size_stats;
> -struct proc_maps_private {
> - struct inode *inode;
> - struct task_struct *task;
> +
> +struct proc_maps_locking_ctx {
> struct mm_struct *mm;
> - struct vma_iterator iter;
> - loff_t last_pos;
> #ifdef CONFIG_PER_VMA_LOCK
> bool mmap_locked;
> struct vm_area_struct *locked_vma;
> #endif
> +};
> +
> +struct proc_maps_private {
> + struct inode *inode;
> + struct task_struct *task;
> + struct vma_iterator iter;
> + loff_t last_pos;
> + struct proc_maps_locking_ctx lock_ctx;
> #ifdef CONFIG_NUMA
> struct mempolicy *task_mempolicy;
> #endif
> diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
> index 29cca0e6d0ff..c0968d293b61 100644
> --- a/fs/proc/task_mmu.c
> +++ b/fs/proc/task_mmu.c
> @@ -132,18 +132,18 @@ static void release_task_mempolicy(struct proc_maps_private *priv)
>
> #ifdef CONFIG_PER_VMA_LOCK
>
> -static void unlock_vma(struct proc_maps_private *priv)
> +static void unlock_ctx_vma(struct proc_maps_locking_ctx *lock_ctx)
> {
> - if (priv->locked_vma) {
> - vma_end_read(priv->locked_vma);
> - priv->locked_vma = NULL;
> + if (lock_ctx->locked_vma) {
> + vma_end_read(lock_ctx->locked_vma);
> + lock_ctx->locked_vma = NULL;
> }
> }
>
> static const struct seq_operations proc_pid_maps_op;
>
> static inline bool lock_vma_range(struct seq_file *m,
> - struct proc_maps_private *priv)
> + struct proc_maps_locking_ctx *lock_ctx)
> {
> /*
> * smaps and numa_maps perform page table walk, therefore require
> @@ -151,25 +151,25 @@ static inline bool lock_vma_range(struct seq_file *m,
> * walking the vma tree under rcu read protection.
> */
> if (m->op != &proc_pid_maps_op) {
> - if (mmap_read_lock_killable(priv->mm))
> + if (mmap_read_lock_killable(lock_ctx->mm))
> return false;
>
> - priv->mmap_locked = true;
> + lock_ctx->mmap_locked = true;
> } else {
> rcu_read_lock();
> - priv->locked_vma = NULL;
> - priv->mmap_locked = false;
> + lock_ctx->locked_vma = NULL;
> + lock_ctx->mmap_locked = false;
> }
>
> return true;
> }
>
> -static inline void unlock_vma_range(struct proc_maps_private *priv)
> +static inline void unlock_vma_range(struct proc_maps_locking_ctx *lock_ctx)
> {
> - if (priv->mmap_locked) {
> - mmap_read_unlock(priv->mm);
> + if (lock_ctx->mmap_locked) {
> + mmap_read_unlock(lock_ctx->mm);
> } else {
> - unlock_vma(priv);
> + unlock_ctx_vma(lock_ctx);
> rcu_read_unlock();
> }
> }
> @@ -177,15 +177,16 @@ static inline void unlock_vma_range(struct proc_maps_private *priv)
> static struct vm_area_struct *get_next_vma(struct proc_maps_private *priv,
> loff_t last_pos)
> {
> + struct proc_maps_locking_ctx *lock_ctx = &priv->lock_ctx;
> struct vm_area_struct *vma;
>
> - if (priv->mmap_locked)
> + if (lock_ctx->mmap_locked)
> return vma_next(&priv->iter);
>
> - unlock_vma(priv);
> - vma = lock_next_vma(priv->mm, &priv->iter, last_pos);
> + unlock_ctx_vma(lock_ctx);
> + vma = lock_next_vma(lock_ctx->mm, &priv->iter, last_pos);
> if (!IS_ERR_OR_NULL(vma))
> - priv->locked_vma = vma;
> + lock_ctx->locked_vma = vma;
>
> return vma;
> }
> @@ -193,14 +194,16 @@ static struct vm_area_struct *get_next_vma(struct proc_maps_private *priv,
> static inline bool fallback_to_mmap_lock(struct proc_maps_private *priv,
> loff_t pos)
> {
> - if (priv->mmap_locked)
> + struct proc_maps_locking_ctx *lock_ctx = &priv->lock_ctx;
> +
> + if (lock_ctx->mmap_locked)
> return false;
>
> rcu_read_unlock();
> - mmap_read_lock(priv->mm);
> + mmap_read_lock(lock_ctx->mm);
> /* Reinitialize the iterator after taking mmap_lock */
> vma_iter_set(&priv->iter, pos);
> - priv->mmap_locked = true;
> + lock_ctx->mmap_locked = true;
>
> return true;
> }
> @@ -208,14 +211,14 @@ static inline bool fallback_to_mmap_lock(struct proc_maps_private *priv,
> #else /* CONFIG_PER_VMA_LOCK */
>
> static inline bool lock_vma_range(struct seq_file *m,
> - struct proc_maps_private *priv)
> + struct proc_maps_locking_ctx *lock_ctx)
> {
> - return mmap_read_lock_killable(priv->mm) == 0;
> + return mmap_read_lock_killable(lock_ctx->mm) == 0;
> }
>
> -static inline void unlock_vma_range(struct proc_maps_private *priv)
> +static inline void unlock_vma_range(struct proc_maps_locking_ctx *lock_ctx)
> {
> - mmap_read_unlock(priv->mm);
> + mmap_read_unlock(lock_ctx->mm);
> }
>
> static struct vm_area_struct *get_next_vma(struct proc_maps_private *priv,
> @@ -258,7 +261,7 @@ static struct vm_area_struct *proc_get_vma(struct seq_file *m, loff_t *ppos)
> *ppos = vma->vm_end;
> } else {
> *ppos = SENTINEL_VMA_GATE;
> - vma = get_gate_vma(priv->mm);
> + vma = get_gate_vma(priv->lock_ctx.mm);
> }
>
> return vma;
> @@ -267,6 +270,7 @@ static struct vm_area_struct *proc_get_vma(struct seq_file *m, loff_t *ppos)
> static void *m_start(struct seq_file *m, loff_t *ppos)
> {
> struct proc_maps_private *priv = m->private;
> + struct proc_maps_locking_ctx *lock_ctx;
> loff_t last_addr = *ppos;
> struct mm_struct *mm;
>
> @@ -278,14 +282,15 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
> if (!priv->task)
> return ERR_PTR(-ESRCH);
>
> - mm = priv->mm;
> + lock_ctx = &priv->lock_ctx;
> + mm = lock_ctx->mm;
> if (!mm || !mmget_not_zero(mm)) {
> put_task_struct(priv->task);
> priv->task = NULL;
> return NULL;
> }
>
> - if (!lock_vma_range(m, priv)) {
> + if (!lock_vma_range(m, lock_ctx)) {
> mmput(mm);
> put_task_struct(priv->task);
> priv->task = NULL;
> @@ -318,13 +323,13 @@ static void *m_next(struct seq_file *m, void *v, loff_t *ppos)
> static void m_stop(struct seq_file *m, void *v)
> {
> struct proc_maps_private *priv = m->private;
> - struct mm_struct *mm = priv->mm;
> + struct mm_struct *mm = priv->lock_ctx.mm;
>
> if (!priv->task)
> return;
>
> release_task_mempolicy(priv);
> - unlock_vma_range(priv);
> + unlock_vma_range(&priv->lock_ctx);
> mmput(mm);
> put_task_struct(priv->task);
> priv->task = NULL;
> @@ -339,9 +344,9 @@ static int proc_maps_open(struct inode *inode, struct file *file,
> return -ENOMEM;
>
> priv->inode = inode;
> - priv->mm = proc_mem_open(inode, PTRACE_MODE_READ);
> - if (IS_ERR(priv->mm)) {
> - int err = PTR_ERR(priv->mm);
> + priv->lock_ctx.mm = proc_mem_open(inode, PTRACE_MODE_READ);
> + if (IS_ERR(priv->lock_ctx.mm)) {
> + int err = PTR_ERR(priv->lock_ctx.mm);
>
> seq_release_private(inode, file);
> return err;
> @@ -355,8 +360,8 @@ static int proc_map_release(struct inode *inode, struct file *file)
> struct seq_file *seq = file->private_data;
> struct proc_maps_private *priv = seq->private;
>
> - if (priv->mm)
> - mmdrop(priv->mm);
> + if (priv->lock_ctx.mm)
> + mmdrop(priv->lock_ctx.mm);
>
> return seq_release_private(inode, file);
> }
> @@ -610,7 +615,7 @@ static int do_procmap_query(struct proc_maps_private *priv, void __user *uarg)
> if (!!karg.build_id_size != !!karg.build_id_addr)
> return -EINVAL;
>
> - mm = priv->mm;
> + mm = priv->lock_ctx.mm;
> if (!mm || !mmget_not_zero(mm))
> return -ESRCH;
>
> @@ -1311,7 +1316,7 @@ static int show_smaps_rollup(struct seq_file *m, void *v)
> {
> struct proc_maps_private *priv = m->private;
> struct mem_size_stats mss = {};
> - struct mm_struct *mm = priv->mm;
> + struct mm_struct *mm = priv->lock_ctx.mm;
> struct vm_area_struct *vma;
> unsigned long vma_start = 0, last_vma_end = 0;
> int ret = 0;
> @@ -1456,9 +1461,9 @@ static int smaps_rollup_open(struct inode *inode, struct file *file)
> goto out_free;
>
> priv->inode = inode;
> - priv->mm = proc_mem_open(inode, PTRACE_MODE_READ);
> - if (IS_ERR_OR_NULL(priv->mm)) {
> - ret = priv->mm ? PTR_ERR(priv->mm) : -ESRCH;
> + priv->lock_ctx.mm = proc_mem_open(inode, PTRACE_MODE_READ);
> + if (IS_ERR_OR_NULL(priv->lock_ctx.mm)) {
> + ret = priv->lock_ctx.mm ? PTR_ERR(priv->lock_ctx.mm) : -ESRCH;
>
> single_release(inode, file);
> goto out_free;
> @@ -1476,8 +1481,8 @@ static int smaps_rollup_release(struct inode *inode, struct file *file)
> struct seq_file *seq = file->private_data;
> struct proc_maps_private *priv = seq->private;
>
> - if (priv->mm)
> - mmdrop(priv->mm);
> + if (priv->lock_ctx.mm)
> + mmdrop(priv->lock_ctx.mm);
>
> kfree(priv);
> return single_release(inode, file);
> diff --git a/fs/proc/task_nommu.c b/fs/proc/task_nommu.c
> index 59bfd61d653a..d362919f4f68 100644
> --- a/fs/proc/task_nommu.c
> +++ b/fs/proc/task_nommu.c
> @@ -204,7 +204,7 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
> if (!priv->task)
> return ERR_PTR(-ESRCH);
>
> - mm = priv->mm;
> + mm = priv->lock_ctx.mm;
> if (!mm || !mmget_not_zero(mm)) {
> put_task_struct(priv->task);
> priv->task = NULL;
> @@ -226,7 +226,7 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
> static void m_stop(struct seq_file *m, void *v)
> {
> struct proc_maps_private *priv = m->private;
> - struct mm_struct *mm = priv->mm;
> + struct mm_struct *mm = priv->lock_ctx.mm;
>
> if (!priv->task)
> return;
> @@ -259,9 +259,9 @@ static int maps_open(struct inode *inode, struct file *file,
> return -ENOMEM;
>
> priv->inode = inode;
> - priv->mm = proc_mem_open(inode, PTRACE_MODE_READ);
> - if (IS_ERR_OR_NULL(priv->mm)) {
> - int err = priv->mm ? PTR_ERR(priv->mm) : -ESRCH;
> + priv->lock_ctx.mm = proc_mem_open(inode, PTRACE_MODE_READ);
> + if (IS_ERR_OR_NULL(priv->lock_ctx.mm)) {
> + int err = priv->lock_ctx.mm ? PTR_ERR(priv->lock_ctx.mm) : -ESRCH;
We could abstract out lock_ctx here also, but I'm not going to be picky about
it.
>
> seq_release_private(inode, file);
> return err;
> @@ -276,8 +276,8 @@ static int map_release(struct inode *inode, struct file *file)
> struct seq_file *seq = file->private_data;
> struct proc_maps_private *priv = seq->private;
>
> - if (priv->mm)
> - mmdrop(priv->mm);
> + if (priv->lock_ctx.mm)
> + mmdrop(priv->lock_ctx.mm);
>
> return seq_release_private(inode, file);
> }
> --
> 2.50.1.703.g449372360f-goog
>
Powered by blists - more mailing lists