[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <6bebe582-705c-2b7a-3286-1c86a15619f2@molgen.mpg.de>
Date: Mon, 22 May 2023 17:05:28 +0200
From: Paul Menzel <pmenzel@...gen.mpg.de>
To: Alexander Lobakin <aleksander.lobakin@...el.com>
Cc: "David S. Miller" <davem@...emloft.net>,
Eric Dumazet <edumazet@...gle.com>,
Jakub Kicinski <kuba@...nel.org>,
Paolo Abeni <pabeni@...hat.com>,
Jesper Dangaard Brouer <hawk@...nel.org>,
Larysa Zaremba <larysa.zaremba@...el.com>,
netdev@...r.kernel.org,
Ilias Apalodimas <ilias.apalodimas@...aro.org>,
linux-kernel@...r.kernel.org,
Michal Kubiak <michal.kubiak@...el.com>,
intel-wired-lan@...ts.osuosl.org, Christoph Hellwig <hch@....de>,
Magnus Karlsson <magnus.karlsson@...el.com>
Subject: Re: [Intel-wired-lan] [PATCH net-next 10/11] libie: add per-queue
Page Pool stats
Dear Alexander,
Thank you for your patch.
Am 16.05.23 um 18:18 schrieb Alexander Lobakin:
> Expand the libie generic per-queue stats with the generic Page Pool
> stats provided by the API itself, when CONFIG_PAGE_POOL is enable.
enable*d*
> When it's not, there'll be no such fields in the stats structure, so
> no space wasted.
> They are also a bit special in terms of how they are obtained. One
> &page_pool accumulates statistics until it's destroyed obviously,
> which happens on ifdown. So, in order to not lose any statistics,
> get the stats and store in the queue container before destroying
> a pool. This container survives ifups/downs, so it basically stores
> the statistics accumulated since the very first pool was allocated
> on this queue. When it's needed to export the stats, first get the
> numbers from this container and then add the "live" numbers -- the
> ones that the current active pool returns. The result values will
> always represent the actual device-lifetime* stats.
> There's a cast from &page_pool_stats to `u64 *` in a couple functions,
> but they are guarded with stats asserts to make sure it's safe to do.
> FWIW it saves a lot of object code.
>
> Signed-off-by: Alexander Lobakin <aleksander.lobakin@...el.com>
> ---
> drivers/net/ethernet/intel/libie/internal.h | 23 +++++++
> drivers/net/ethernet/intel/libie/rx.c | 20 ++++++
> drivers/net/ethernet/intel/libie/stats.c | 72 ++++++++++++++++++++-
> include/linux/net/intel/libie/rx.h | 4 ++
> include/linux/net/intel/libie/stats.h | 39 ++++++++++-
> 5 files changed, 155 insertions(+), 3 deletions(-)
> create mode 100644 drivers/net/ethernet/intel/libie/internal.h
>
> diff --git a/drivers/net/ethernet/intel/libie/internal.h b/drivers/net/ethernet/intel/libie/internal.h
> new file mode 100644
> index 000000000000..083398dc37c6
> --- /dev/null
> +++ b/drivers/net/ethernet/intel/libie/internal.h
> @@ -0,0 +1,23 @@
> +/* SPDX-License-Identifier: GPL-2.0-only */
> +/* libie internal declarations not to be used in drivers.
> + *
> + * Copyright(c) 2023 Intel Corporation.
> + */
> +
> +#ifndef __LIBIE_INTERNAL_H
> +#define __LIBIE_INTERNAL_H
> +
> +struct libie_rq_stats;
> +struct page_pool;
> +
> +#ifdef CONFIG_PAGE_POOL_STATS
> +void libie_rq_stats_sync_pp(struct libie_rq_stats *stats,
> + struct page_pool *pool);
> +#else
> +static inline void libie_rq_stats_sync_pp(struct libie_rq_stats *stats,
> + struct page_pool *pool)
> +{
> +}
> +#endif
> +
> +#endif /* __LIBIE_INTERNAL_H */
> diff --git a/drivers/net/ethernet/intel/libie/rx.c b/drivers/net/ethernet/intel/libie/rx.c
> index d68eab76593c..128f134aca6d 100644
> --- a/drivers/net/ethernet/intel/libie/rx.c
> +++ b/drivers/net/ethernet/intel/libie/rx.c
> @@ -3,6 +3,8 @@
>
> #include <linux/net/intel/libie/rx.h>
>
> +#include "internal.h"
> +
> /* O(1) converting i40e/ice/iavf's 8/10-bit hardware packet type to a parsed
> * bitfield struct.
> */
> @@ -133,6 +135,24 @@ struct page_pool *libie_rx_page_pool_create(struct napi_struct *napi,
> }
> EXPORT_SYMBOL_NS_GPL(libie_rx_page_pool_create, LIBIE);
>
> +/**
> + * libie_rx_page_pool_destroy - destroy a &page_pool created by libie
> + * @pool: pool to destroy
> + * @stats: RQ stats from the ring (or %NULL to skip updating PP stats)
> + *
> + * As the stats usually has the same lifetime as the device, but PP is usually
> + * created/destroyed on ifup/ifdown, in order to not lose the stats accumulated
> + * during the last ifup, the PP stats need to be added to the driver stats
> + * container. Then the PP gets destroyed.
> + */
> +void libie_rx_page_pool_destroy(struct page_pool *pool,
> + struct libie_rq_stats *stats)
> +{
> + libie_rq_stats_sync_pp(stats, pool);
> + page_pool_destroy(pool);
> +}
> +EXPORT_SYMBOL_NS_GPL(libie_rx_page_pool_destroy, LIBIE);
> +
> MODULE_AUTHOR("Intel Corporation");
> MODULE_DESCRIPTION("Intel(R) Ethernet common library");
> MODULE_LICENSE("GPL");
> diff --git a/drivers/net/ethernet/intel/libie/stats.c b/drivers/net/ethernet/intel/libie/stats.c
> index 61456842a362..95bbb38c39e3 100644
> --- a/drivers/net/ethernet/intel/libie/stats.c
> +++ b/drivers/net/ethernet/intel/libie/stats.c
> @@ -4,6 +4,8 @@
> #include <linux/ethtool.h>
> #include <linux/net/intel/libie/stats.h>
>
> +#include "internal.h"
> +
> /* Rx per-queue stats */
>
> static const char * const libie_rq_stats_str[] = {
> @@ -14,6 +16,70 @@ static const char * const libie_rq_stats_str[] = {
>
> #define LIBIE_RQ_STATS_NUM ARRAY_SIZE(libie_rq_stats_str)
>
> +#ifdef CONFIG_PAGE_POOL_STATS
> +/**
> + * libie_rq_stats_get_pp - get the current stats from a &page_pool
> + * @sarr: local array to add stats to
> + * @pool: pool to get the stats from
> + *
> + * Adds the current "live" stats from an online PP to the stats read from
> + * the RQ container, so that the actual totals will be returned.
> + */
> +static void libie_rq_stats_get_pp(u64 *sarr, struct page_pool *pool)
> +{
> + struct page_pool_stats *pps;
> + /* Used only to calculate pos below */
> + struct libie_rq_stats tmp;
> + u32 pos;
> +
> + /* Validate the libie PP stats array can be casted <-> PP struct */
> + static_assert(sizeof(tmp.pp) == sizeof(*pps));
> +
> + if (!pool)
> + return;
> +
> + /* Position of the first Page Pool stats field */
> + pos = (u64_stats_t *)&tmp.pp - tmp.raw;
> + pps = (typeof(pps))&sarr[pos];
> +
> + page_pool_get_stats(pool, pps);
> +}
> +
> +/**
> + * libie_rq_stats_sync_pp - add the current PP stats to the RQ stats container
> + * @stats: stats structure to update
> + * @pool: pool to read the stats
> + *
> + * Called by libie_rx_page_pool_destroy() to save the stats before destroying
> + * the pool.
> + */
> +void libie_rq_stats_sync_pp(struct libie_rq_stats *stats,
> + struct page_pool *pool)
> +{
> + u64_stats_t *qarr = (u64_stats_t *)&stats->pp;
> + struct page_pool_stats pps = { };
> + u64 *sarr = (u64 *)&pps;
> +
> + if (!stats)
> + return;
> +
> + page_pool_get_stats(pool, &pps);
> +
> + u64_stats_update_begin(&stats->syncp);
> +
> + for (u32 i = 0; i < sizeof(pps) / sizeof(*sarr); i++)
> + u64_stats_add(&qarr[i], sarr[i]);
> +
> + u64_stats_update_end(&stats->syncp);
> +}
> +#else
> +static inline void libie_rq_stats_get_pp(u64 *sarr, struct page_pool *pool)
> +{
> +}
> +
> +/* static inline void libie_rq_stats_sync_pp() is declared in "internal.h" */
> +#endif
> +
> /**
> * libie_rq_stats_get_sset_count - get the number of Ethtool RQ stats provided
> *
> @@ -41,8 +107,10 @@ EXPORT_SYMBOL_NS_GPL(libie_rq_stats_get_strings, LIBIE);
> * libie_rq_stats_get_data - get the RQ stats in Ethtool format
> * @data: reference to the cursor pointing to the output array
> * @stats: RQ stats container from the queue
> + * @pool: &page_pool from the queue (%NULL to ignore PP "live" stats)
> */
> -void libie_rq_stats_get_data(u64 **data, const struct libie_rq_stats *stats)
> +void libie_rq_stats_get_data(u64 **data, const struct libie_rq_stats *stats,
> + struct page_pool *pool)
> {
> u64 sarr[LIBIE_RQ_STATS_NUM];
> u32 start;
> @@ -54,6 +122,8 @@ void libie_rq_stats_get_data(u64 **data, const struct libie_rq_stats *stats)
> sarr[i] = u64_stats_read(&stats->raw[i]);
> } while (u64_stats_fetch_retry(&stats->syncp, start));
>
> + libie_rq_stats_get_pp(sarr, pool);
> +
> for (u32 i = 0; i < LIBIE_RQ_STATS_NUM; i++)
> (*data)[i] += sarr[i];
>
> diff --git a/include/linux/net/intel/libie/rx.h b/include/linux/net/intel/libie/rx.h
> index f6ba3b19b7e2..474ffd689001 100644
> --- a/include/linux/net/intel/libie/rx.h
> +++ b/include/linux/net/intel/libie/rx.h
> @@ -160,7 +160,11 @@ static inline void libie_skb_set_hash(struct sk_buff *skb, u32 hash,
> /* Maximum frame size minus LL overhead */
> #define LIBIE_MAX_MTU (LIBIE_MAX_RX_FRM_LEN - LIBIE_RX_LL_LEN)
>
> +struct libie_rq_stats;
> +
> struct page_pool *libie_rx_page_pool_create(struct napi_struct *napi,
> u32 size);
> +void libie_rx_page_pool_destroy(struct page_pool *pool,
> + struct libie_rq_stats *stats);
>
> #endif /* __LIBIE_RX_H */
> diff --git a/include/linux/net/intel/libie/stats.h b/include/linux/net/intel/libie/stats.h
> index dbbc98bbd3a7..23ca0079a905 100644
> --- a/include/linux/net/intel/libie/stats.h
> +++ b/include/linux/net/intel/libie/stats.h
> @@ -49,6 +49,17 @@
> * fragments: number of processed descriptors carrying only a fragment
> * alloc_page_fail: number of Rx page allocation fails
> * build_skb_fail: number of build_skb() fails
> + * pp_alloc_fast: pages taken from the cache or ring
> + * pp_alloc_slow: actual page allocations
> + * pp_alloc_slow_ho: non-order-0 page allocations
> + * pp_alloc_empty: number of times the pool was empty
> + * pp_alloc_refill: number of cache refills
> + * pp_alloc_waive: NUMA node mismatches during recycling
> + * pp_recycle_cached: direct recyclings into the cache
> + * pp_recycle_cache_full: number of times the cache was full
> + * pp_recycle_ring: recyclings into the ring
> + * pp_recycle_ring_full: number of times the ring was full
> + * pp_recycle_released_ref: pages released due to elevated refcnt
> */
>
> #define DECLARE_LIBIE_RQ_NAPI_STATS(act) \
> @@ -60,9 +71,29 @@
> act(alloc_page_fail) \
> act(build_skb_fail)
>
> +#ifdef CONFIG_PAGE_POOL_STATS
> +#define DECLARE_LIBIE_RQ_PP_STATS(act) \
> + act(pp_alloc_fast) \
> + act(pp_alloc_slow) \
> + act(pp_alloc_slow_ho) \
> + act(pp_alloc_empty) \
> + act(pp_alloc_refill) \
> + act(pp_alloc_waive) \
> + act(pp_recycle_cached) \
> + act(pp_recycle_cache_full) \
> + act(pp_recycle_ring) \
> + act(pp_recycle_ring_full) \
> + act(pp_recycle_released_ref)
> +#else
> +#define DECLARE_LIBIE_RQ_PP_STATS(act)
> +#endif
> +
> #define DECLARE_LIBIE_RQ_STATS(act) \
> DECLARE_LIBIE_RQ_NAPI_STATS(act) \
> - DECLARE_LIBIE_RQ_FAIL_STATS(act)
> + DECLARE_LIBIE_RQ_FAIL_STATS(act) \
> + DECLARE_LIBIE_RQ_PP_STATS(act)
> +
> +struct page_pool;
>
> struct libie_rq_stats {
> struct u64_stats_sync syncp;
> @@ -72,6 +103,9 @@ struct libie_rq_stats {
> #define act(s) u64_stats_t s;
> DECLARE_LIBIE_RQ_NAPI_STATS(act);
> DECLARE_LIBIE_RQ_FAIL_STATS(act);
> + struct_group(pp,
> + DECLARE_LIBIE_RQ_PP_STATS(act);
> + );
> #undef act
> };
> DECLARE_FLEX_ARRAY(u64_stats_t, raw);
> @@ -110,7 +144,8 @@ libie_rq_napi_stats_add(struct libie_rq_stats *qs,
>
> u32 libie_rq_stats_get_sset_count(void);
> void libie_rq_stats_get_strings(u8 **data, u32 qid);
> -void libie_rq_stats_get_data(u64 **data, const struct libie_rq_stats *stats);
> +void libie_rq_stats_get_data(u64 **data, const struct libie_rq_stats *stats,
> + struct page_pool *pool);
>
> /* Tx per-queue stats:
> * packets: packets sent from this queue
Reviewed-by: Paul Menzel <pmenzel@...gen.mpg.de>
Kind regards,
Paul
Powered by blists - more mailing lists