[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20230111042214.907030-3-willy@infradead.org>
Date: Wed, 11 Jan 2023 04:21:50 +0000
From: "Matthew Wilcox (Oracle)" <willy@...radead.org>
To: Jesper Dangaard Brouer <hawk@...nel.org>,
Ilias Apalodimas <ilias.apalodimas@...aro.org>
Cc: "Matthew Wilcox (Oracle)" <willy@...radead.org>,
netdev@...r.kernel.org, linux-mm@...ck.org,
Shakeel Butt <shakeelb@...gle.com>,
Jesper Dangaard Brouer <brouer@...hat.com>,
Jesse Brandeburg <jesse.brandeburg@...el.com>
Subject: [PATCH v3 02/26] netmem: Add utility functions
netmem_page() is defined this way to preserve constness. page_netmem()
doesn't call compound_head() because netmem users always use the head
page; it does include a debugging assert to check that it's true.
Signed-off-by: Matthew Wilcox (Oracle) <willy@...radead.org>
Acked-by: Jesper Dangaard Brouer <brouer@...hat.com>
Reviewed-by: Jesse Brandeburg <jesse.brandeburg@...el.com>
---
include/net/page_pool.h | 59 +++++++++++++++++++++++++++++++++++++++++
1 file changed, 59 insertions(+)
diff --git a/include/net/page_pool.h b/include/net/page_pool.h
index cbea4df54918..414907e67ff6 100644
--- a/include/net/page_pool.h
+++ b/include/net/page_pool.h
@@ -96,6 +96,65 @@ NETMEM_MATCH(_refcount, _refcount);
#undef NETMEM_MATCH
static_assert(sizeof(struct netmem) <= sizeof(struct page));
+#define netmem_page(nmem) (_Generic((nmem), \
+ const struct netmem *: (const struct page *)nmem, \
+ struct netmem *: (struct page *)nmem))
+
+static inline struct netmem *page_netmem(struct page *page)
+{
+ VM_BUG_ON_PAGE(PageTail(page), page);
+ return (struct netmem *)page;
+}
+
+static inline unsigned long netmem_pfn(const struct netmem *nmem)
+{
+ return page_to_pfn(netmem_page(nmem));
+}
+
+static inline unsigned long netmem_nid(const struct netmem *nmem)
+{
+ return page_to_nid(netmem_page(nmem));
+}
+
+static inline struct netmem *virt_to_netmem(const void *x)
+{
+ return page_netmem(virt_to_head_page(x));
+}
+
+static inline void *netmem_to_virt(const struct netmem *nmem)
+{
+ return page_to_virt(netmem_page(nmem));
+}
+
+static inline void *netmem_address(const struct netmem *nmem)
+{
+ return page_address(netmem_page(nmem));
+}
+
+static inline int netmem_ref_count(const struct netmem *nmem)
+{
+ return page_ref_count(netmem_page(nmem));
+}
+
+static inline void netmem_get(struct netmem *nmem)
+{
+ struct folio *folio = (struct folio *)nmem;
+
+ folio_get(folio);
+}
+
+static inline void netmem_put(struct netmem *nmem)
+{
+ struct folio *folio = (struct folio *)nmem;
+
+ folio_put(folio);
+}
+
+static inline bool netmem_is_pfmemalloc(const struct netmem *nmem)
+{
+ return nmem->pp_magic & BIT(1);
+}
+
/*
* Fast allocation side cache array/stack
*
--
2.35.1
Powered by blists - more mailing lists