[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <176169810678.1424854.571040440352208082.stgit@frogsfrogsfrogs>
Date: Tue, 28 Oct 2025 17:48:52 -0700
From: "Darrick J. Wong" <djwong@...nel.org>
To: djwong@...nel.org, miklos@...redi.hu
Cc: joannelkoong@...il.com, bernd@...ernd.com, neal@...pa.dev,
linux-ext4@...r.kernel.org, linux-fsdevel@...r.kernel.org
Subject: [PATCH 15/31] fuse_trace: implement buffered IO with iomap
From: Darrick J. Wong <djwong@...nel.org>
Add tracepoints for the previous patch.
Signed-off-by: "Darrick J. Wong" <djwong@...nel.org>
---
fs/fuse/fuse_trace.h | 252 ++++++++++++++++++++++++++++++++++++++++++++++++++
fs/fuse/file_iomap.c | 40 ++++++++
2 files changed, 288 insertions(+), 4 deletions(-)
diff --git a/fs/fuse/fuse_trace.h b/fs/fuse/fuse_trace.h
index bd88c46b447997..a9ccb6a7491fc1 100644
--- a/fs/fuse/fuse_trace.h
+++ b/fs/fuse/fuse_trace.h
@@ -224,6 +224,9 @@ DEFINE_FUSE_BACKING_EVENT(fuse_backing_close);
#endif /* CONFIG_FUSE_BACKING */
#if IS_ENABLED(CONFIG_FUSE_IOMAP)
+struct iomap_writepage_ctx;
+struct iomap_ioend;
+
/* tracepoint boilerplate so we don't have to keep doing this */
#define FUSE_IOMAP_OPFLAGS_FIELD \
__field(unsigned, opflags)
@@ -291,7 +294,8 @@ DEFINE_FUSE_BACKING_EVENT(fuse_backing_close);
{ FUSE_IOMAP_OP_UNSHARE, "unshare" }, \
{ FUSE_IOMAP_OP_DAX, "fsdax" }, \
{ FUSE_IOMAP_OP_ATOMIC, "atomic" }, \
- { FUSE_IOMAP_OP_DONTCACHE, "dontcache" }
+ { FUSE_IOMAP_OP_DONTCACHE, "dontcache" }, \
+ { FUSE_IOMAP_OP_WRITEBACK, "writeback" }
#define FUSE_IOMAP_TYPE_STRINGS \
{ FUSE_IOMAP_TYPE_PURE_OVERWRITE, "overwrite" }, \
@@ -306,7 +310,8 @@ DEFINE_FUSE_BACKING_EVENT(fuse_backing_close);
{ FUSE_IOMAP_IOEND_UNWRITTEN, "unwritten" }, \
{ FUSE_IOMAP_IOEND_BOUNDARY, "boundary" }, \
{ FUSE_IOMAP_IOEND_DIRECT, "direct" }, \
- { FUSE_IOMAP_IOEND_APPEND, "append" }
+ { FUSE_IOMAP_IOEND_APPEND, "append" }, \
+ { FUSE_IOMAP_IOEND_WRITEBACK, "writeback" }
#define IOMAP_DIOEND_STRINGS \
{ IOMAP_DIO_UNWRITTEN, "unwritten" }, \
@@ -331,6 +336,12 @@ TRACE_DEFINE_ENUM(FUSE_I_IOMAP);
{ 1 << FUSE_I_EXCLUSIVE, "excl" }, \
{ 1 << FUSE_I_IOMAP, "iomap" }
+#define IOMAP_IOEND_STRINGS \
+ { IOMAP_IOEND_SHARED, "shared" }, \
+ { IOMAP_IOEND_UNWRITTEN, "unwritten" }, \
+ { IOMAP_IOEND_BOUNDARY, "boundary" }, \
+ { IOMAP_IOEND_DIRECT, "direct" }
+
DECLARE_EVENT_CLASS(fuse_iomap_check_class,
TP_PROTO(const char *func, int line, const char *condition),
@@ -670,6 +681,9 @@ DEFINE_EVENT(fuse_iomap_file_io_class, name, \
TP_ARGS(iocb, iter))
DEFINE_FUSE_IOMAP_FILE_IO_EVENT(fuse_iomap_direct_read);
DEFINE_FUSE_IOMAP_FILE_IO_EVENT(fuse_iomap_direct_write);
+DEFINE_FUSE_IOMAP_FILE_IO_EVENT(fuse_iomap_buffered_read);
+DEFINE_FUSE_IOMAP_FILE_IO_EVENT(fuse_iomap_buffered_write);
+DEFINE_FUSE_IOMAP_FILE_IO_EVENT(fuse_iomap_write_zero_eof);
DECLARE_EVENT_CLASS(fuse_iomap_file_ioend_class,
TP_PROTO(const struct kiocb *iocb, const struct iov_iter *iter,
@@ -696,6 +710,8 @@ DEFINE_EVENT(fuse_iomap_file_ioend_class, name, \
TP_ARGS(iocb, iter, ret))
DEFINE_FUSE_IOMAP_FILE_IOEND_EVENT(fuse_iomap_direct_read_end);
DEFINE_FUSE_IOMAP_FILE_IOEND_EVENT(fuse_iomap_direct_write_end);
+DEFINE_FUSE_IOMAP_FILE_IOEND_EVENT(fuse_iomap_buffered_read_end);
+DEFINE_FUSE_IOMAP_FILE_IOEND_EVENT(fuse_iomap_buffered_write_end);
TRACE_EVENT(fuse_iomap_dio_write_end_io,
TP_PROTO(const struct inode *inode, loff_t pos, ssize_t written,
@@ -722,6 +738,238 @@ TRACE_EVENT(fuse_iomap_dio_write_end_io,
__print_flags(__entry->dioendflags, "|", IOMAP_DIOEND_STRINGS),
__entry->error)
);
+
+TRACE_EVENT(fuse_iomap_end_ioend,
+ TP_PROTO(const struct iomap_ioend *ioend),
+
+ TP_ARGS(ioend),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ __field(unsigned int, ioendflags)
+ __field(int, error)
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(ioend->io_inode, fi, fm);
+ __entry->offset = ioend->io_offset;
+ __entry->length = ioend->io_size;
+ __entry->ioendflags = ioend->io_flags;
+ __entry->error = blk_status_to_errno(ioend->io_bio.bi_status);
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT() " ioendflags (%s) error %d",
+ FUSE_IO_RANGE_PRINTK_ARGS(),
+ __print_flags(__entry->ioendflags, "|", IOMAP_IOEND_STRINGS),
+ __entry->error)
+);
+
+TRACE_EVENT(fuse_iomap_writeback_range,
+ TP_PROTO(const struct inode *inode, u64 offset, unsigned int count,
+ u64 end_pos),
+
+ TP_ARGS(inode, offset, count, end_pos),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ __field(uint64_t, end_pos)
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(inode, fi, fm);
+ __entry->offset = offset;
+ __entry->length = count;
+ __entry->end_pos = end_pos;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT() " end_pos 0x%llx",
+ FUSE_IO_RANGE_PRINTK_ARGS(),
+ __entry->end_pos)
+);
+
+TRACE_EVENT(fuse_iomap_writeback_submit,
+ TP_PROTO(const struct iomap_writepage_ctx *wpc, int error),
+
+ TP_ARGS(wpc, error),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ __field(unsigned int, nr_folios)
+ __field(uint64_t, addr)
+ __field(int, error)
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(wpc->inode, fi, fm);
+ __entry->nr_folios = wpc->nr_folios;
+ __entry->offset = wpc->iomap.offset;
+ __entry->length = wpc->iomap.length;
+ __entry->addr = wpc->iomap.addr << 9;
+ __entry->error = error;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT() " addr 0x%llx nr_folios %u error %d",
+ FUSE_IO_RANGE_PRINTK_ARGS(),
+ __entry->addr,
+ __entry->nr_folios,
+ __entry->error)
+);
+
+TRACE_EVENT(fuse_iomap_discard_folio,
+ TP_PROTO(const struct inode *inode, loff_t offset, size_t count),
+
+ TP_ARGS(inode, offset, count),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(inode, fi, fm);
+ __entry->offset = offset;
+ __entry->length = count;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT(),
+ FUSE_IO_RANGE_PRINTK_ARGS())
+);
+
+TRACE_EVENT(fuse_iomap_writepages,
+ TP_PROTO(const struct inode *inode, const struct writeback_control *wbc),
+
+ TP_ARGS(inode, wbc),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ __field(long, nr_to_write)
+ __field(bool, sync_all)
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(inode, fi, fm);
+ __entry->offset = wbc->range_start;
+ __entry->length = wbc->range_end - wbc->range_start + 1;
+ __entry->nr_to_write = wbc->nr_to_write;
+ __entry->sync_all = wbc->sync_mode == WB_SYNC_ALL;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT() " nr_folios %ld sync_all? %d",
+ FUSE_IO_RANGE_PRINTK_ARGS(),
+ __entry->nr_to_write,
+ __entry->sync_all)
+);
+
+TRACE_EVENT(fuse_iomap_read_folio,
+ TP_PROTO(const struct folio *folio),
+
+ TP_ARGS(folio),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(folio->mapping->host, fi, fm);
+ __entry->offset = folio_pos(folio);
+ __entry->length = folio_size(folio);
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT(),
+ FUSE_IO_RANGE_PRINTK_ARGS())
+);
+
+TRACE_EVENT(fuse_iomap_readahead,
+ TP_PROTO(const struct readahead_control *rac),
+
+ TP_ARGS(rac),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ ),
+
+ TP_fast_assign(
+ struct readahead_control *mutrac = (struct readahead_control *)rac;
+ FUSE_INODE_ASSIGN(file_inode(rac->file), fi, fm);
+ __entry->offset = readahead_pos(mutrac);
+ __entry->length = readahead_length(mutrac);
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT(),
+ FUSE_IO_RANGE_PRINTK_ARGS())
+);
+
+TRACE_EVENT(fuse_iomap_page_mkwrite,
+ TP_PROTO(const struct vm_fault *vmf),
+
+ TP_ARGS(vmf),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ ),
+
+ TP_fast_assign(
+ struct folio *folio = page_folio(vmf->page);
+ FUSE_INODE_ASSIGN(file_inode(vmf->vma->vm_file), fi, fm);
+ __entry->offset = folio_pos(folio);
+ __entry->length = folio_size(folio);
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT(),
+ FUSE_IO_RANGE_PRINTK_ARGS())
+);
+
+DECLARE_EVENT_CLASS(fuse_iomap_file_range_class,
+ TP_PROTO(const struct inode *inode, loff_t offset, loff_t length),
+
+ TP_ARGS(inode, offset, length),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(inode, fi, fm);
+ __entry->offset = offset;
+ __entry->length = length;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT(),
+ FUSE_IO_RANGE_PRINTK_ARGS())
+)
+#define DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(name) \
+DEFINE_EVENT(fuse_iomap_file_range_class, name, \
+ TP_PROTO(const struct inode *inode, loff_t offset, loff_t length), \
+ TP_ARGS(inode, offset, length))
+DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(fuse_iomap_truncate_up);
+DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(fuse_iomap_truncate_down);
+DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(fuse_iomap_punch_range);
+DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(fuse_iomap_setsize);
+DEFINE_FUSE_IOMAP_FILE_RANGE_EVENT(fuse_iomap_flush_unmap_range);
+
+TRACE_EVENT(fuse_iomap_fallocate,
+ TP_PROTO(const struct inode *inode, int mode, loff_t offset,
+ loff_t length, loff_t newsize),
+ TP_ARGS(inode, mode, offset, length, newsize),
+
+ TP_STRUCT__entry(
+ FUSE_IO_RANGE_FIELDS()
+ __field(loff_t, newsize)
+ __field(int, mode)
+ ),
+
+ TP_fast_assign(
+ FUSE_INODE_ASSIGN(inode, fi, fm);
+ __entry->offset = offset;
+ __entry->length = length;
+ __entry->mode = mode;
+ __entry->newsize = newsize;
+ ),
+
+ TP_printk(FUSE_IO_RANGE_FMT() " mode 0x%x newsize 0x%llx",
+ FUSE_IO_RANGE_PRINTK_ARGS(),
+ __entry->mode,
+ __entry->newsize)
+);
#endif /* CONFIG_FUSE_IOMAP */
#endif /* _TRACE_FUSE_H */
diff --git a/fs/fuse/file_iomap.c b/fs/fuse/file_iomap.c
index fd283b98d5e800..897a07f197c797 100644
--- a/fs/fuse/file_iomap.c
+++ b/fs/fuse/file_iomap.c
@@ -1026,6 +1026,8 @@ fuse_iomap_write_zero_eof(
return 1;
}
+ trace_fuse_iomap_write_zero_eof(iocb, from);
+
filemap_invalidate_lock(mapping);
error = fuse_iomap_zero_range(inode, isize, iocb->ki_pos - isize, NULL);
filemap_invalidate_unlock(mapping);
@@ -1138,6 +1140,8 @@ static void fuse_iomap_end_ioend(struct iomap_ioend *ioend)
if (fuse_is_bad(inode))
return;
+ trace_fuse_iomap_end_ioend(ioend);
+
if (ioend->io_flags & IOMAP_IOEND_SHARED)
ioendflags |= FUSE_IOMAP_IOEND_SHARED;
if (ioend->io_flags & IOMAP_IOEND_UNWRITTEN)
@@ -1246,6 +1250,8 @@ static void fuse_iomap_discard_folio(struct folio *folio, loff_t pos, int error)
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_discard_folio(inode, pos, folio_size(folio));
+
printk_ratelimited(KERN_ERR
"page discard on page %px, inode 0x%llx, pos %llu.",
folio, fi->orig_ino, pos);
@@ -1269,6 +1275,8 @@ static ssize_t fuse_iomap_writeback_range(struct iomap_writepage_ctx *wpc,
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_writeback_range(inode, offset, len, end_pos);
+
if (!fuse_iomap_revalidate_writeback(wpc, offset)) {
ret = fuse_iomap_begin(inode, offset, len,
FUSE_IOMAP_OP_WRITEBACK,
@@ -1306,6 +1314,8 @@ static int fuse_iomap_writeback_submit(struct iomap_writepage_ctx *wpc,
ASSERT(fuse_inode_has_iomap(ioend->io_inode));
+ trace_fuse_iomap_writeback_submit(wpc, error);
+
/* always call our ioend function, even if we cancel the bio */
ioend->io_bio.bi_end_io = fuse_iomap_end_bio;
return iomap_ioend_writeback_submit(wpc, error);
@@ -1329,6 +1339,8 @@ static int fuse_iomap_writepages(struct address_space *mapping,
ASSERT(fuse_inode_has_iomap(mapping->host));
+ trace_fuse_iomap_writepages(mapping->host, wbc);
+
return iomap_writepages(&wpc.ctx);
}
@@ -1336,6 +1348,8 @@ static int fuse_iomap_read_folio(struct file *file, struct folio *folio)
{
ASSERT(fuse_inode_has_iomap(file_inode(file)));
+ trace_fuse_iomap_read_folio(folio);
+
return iomap_read_folio(folio, &fuse_iomap_ops);
}
@@ -1343,6 +1357,8 @@ static void fuse_iomap_readahead(struct readahead_control *rac)
{
ASSERT(fuse_inode_has_iomap(file_inode(rac->file)));
+ trace_fuse_iomap_readahead(rac);
+
iomap_readahead(rac, &fuse_iomap_ops);
}
@@ -1391,6 +1407,8 @@ static vm_fault_t fuse_iomap_page_mkwrite(struct vm_fault *vmf)
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_page_mkwrite(vmf);
+
sb_start_pagefault(inode->i_sb);
file_update_time(vmf->vma->vm_file);
@@ -1424,6 +1442,8 @@ ssize_t fuse_iomap_buffered_read(struct kiocb *iocb, struct iov_iter *to)
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_buffered_read(iocb, to);
+
if (!iov_iter_count(to))
return 0; /* skip atime */
@@ -1435,6 +1455,7 @@ ssize_t fuse_iomap_buffered_read(struct kiocb *iocb, struct iov_iter *to)
ret = generic_file_read_iter(iocb, to);
inode_unlock_shared(inode);
+ trace_fuse_iomap_buffered_read_end(iocb, to, ret);
return ret;
}
@@ -1447,6 +1468,8 @@ ssize_t fuse_iomap_buffered_write(struct kiocb *iocb, struct iov_iter *from)
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_buffered_write(iocb, from);
+
if (!iov_iter_count(from))
return 0;
@@ -1475,6 +1498,7 @@ ssize_t fuse_iomap_buffered_write(struct kiocb *iocb, struct iov_iter *from)
/* Handle various SYNC-type writes */
ret = generic_write_sync(iocb, ret);
}
+ trace_fuse_iomap_buffered_write_end(iocb, from, ret);
return ret;
}
@@ -1520,11 +1544,17 @@ fuse_iomap_setsize_start(
* extension, or zeroing out the rest of the block on a downward
* truncate.
*/
- if (newsize > oldsize)
+ if (newsize > oldsize) {
+ trace_fuse_iomap_truncate_up(inode, oldsize, newsize - oldsize);
+
error = fuse_iomap_zero_range(inode, oldsize, newsize - oldsize,
&did_zeroing);
- else
+ } else {
+ trace_fuse_iomap_truncate_down(inode, newsize,
+ oldsize - newsize);
+
error = fuse_iomap_truncate_page(inode, newsize, &did_zeroing);
+ }
if (error)
return error;
@@ -1571,6 +1601,8 @@ int fuse_iomap_flush_unmap_range(struct inode *inode, loff_t pos,
start = round_down(pos, rounding);
end = round_up(endpos + 1, rounding) - 1;
+ trace_fuse_iomap_flush_unmap_range(inode, start, end + 1 - start);
+
error = filemap_write_and_wait_range(inode->i_mapping, start, end);
if (error)
return error;
@@ -1584,6 +1616,8 @@ static int fuse_iomap_punch_range(struct inode *inode, loff_t offset,
loff_t isize = i_size_read(inode);
int error;
+ trace_fuse_iomap_punch_range(inode, offset, length);
+
/*
* Now that we've unmap all full blocks we'll have to zero out any
* partial block at the beginning and/or end. iomap_zero_range is
@@ -1627,6 +1661,8 @@ fuse_iomap_fallocate(
ASSERT(fuse_inode_has_iomap(inode));
+ trace_fuse_iomap_fallocate(inode, mode, offset, length, new_size);
+
/*
* If we unmapped blocks from the file range, then we zero the
* pagecache for those regions and push them to disk rather than make
Powered by blists - more mailing lists