[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200622105923.0000710b@Huawei.com>
Date: Mon, 22 Jun 2020 10:59:23 +0100
From: Jonathan Cameron <Jonathan.Cameron@...wei.com>
To: Barry Song <song.bao.hua@...ilicon.com>
CC: <herbert@...dor.apana.org.au>, <davem@...emloft.net>,
Seth Jennings <sjenning@...hat.com>, <linuxarm@...wei.com>,
<linux-kernel@...r.kernel.org>, <linux-mm@...ck.org>,
<linux-crypto@...r.kernel.org>, <akpm@...ux-foundation.org>,
Dan Streetman <ddstreet@...e.org>,
Vitaly Wool <vitaly.wool@...sulko.com>
Subject: Re: [PATCH 1/3] crypto: permit users to specify numa node of acomp
hardware
On Mon, 22 Jun 2020 14:48:59 +1200
Barry Song <song.bao.hua@...ilicon.com> wrote:
> For a Linux server with NUMA, there are possibly multiple (de)compressors
> which are either local or remote to some NUMA node. Some drivers will
> automatically use the (de)compressor near the CPU calling acomp_alloc().
> However, it is not necessarily correct because users who send acomp_req
> could be from different NUMA node with the CPU which allocates acomp.
>
> Just like kernel has kmalloc() and kmalloc_node(), here crypto can have
> same support.
>
> Cc: Seth Jennings <sjenning@...hat.com>
> Cc: Dan Streetman <ddstreet@...e.org>
> Cc: Vitaly Wool <vitaly.wool@...sulko.com>
> Cc: Andrew Morton <akpm@...ux-foundation.org>
> Signed-off-by: Barry Song <song.bao.hua@...ilicon.com>
Hi Barry,
Seems sensible to me. A few trivial comments inline.
Thanks,
Jonathan
> ---
> crypto/acompress.c | 8 ++++++++
> crypto/api.c | 22 ++++++++++++++--------
> crypto/internal.h | 23 +++++++++++++++++++----
> include/crypto/acompress.h | 7 +++++++
> include/linux/crypto.h | 3 ++-
> 5 files changed, 50 insertions(+), 13 deletions(-)
>
> diff --git a/crypto/acompress.c b/crypto/acompress.c
> index 84a76723e851..c32c72048a1c 100644
> --- a/crypto/acompress.c
> +++ b/crypto/acompress.c
> @@ -109,6 +109,14 @@ struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
> }
> EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
>
> +struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
> + u32 mask, int node)
> +{
> + return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask,
> + node);
> +}
> +EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node);
> +
> struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
> {
> struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
> diff --git a/crypto/api.c b/crypto/api.c
> index edcf690800d4..4ecf712286af 100644
> --- a/crypto/api.c
> +++ b/crypto/api.c
> @@ -433,8 +433,9 @@ struct crypto_tfm *crypto_alloc_base(const char *alg_name, u32 type, u32 mask)
> }
> EXPORT_SYMBOL_GPL(crypto_alloc_base);
>
> -void *crypto_create_tfm(struct crypto_alg *alg,
> - const struct crypto_type *frontend)
> +void *crypto_create_tfm_node(struct crypto_alg *alg,
> + const struct crypto_type *frontend,
> + int node)
> {
> char *mem;
> struct crypto_tfm *tfm = NULL;
> @@ -451,6 +452,7 @@ void *crypto_create_tfm(struct crypto_alg *alg,
>
> tfm = (struct crypto_tfm *)(mem + tfmsize);
> tfm->__crt_alg = alg;
> + tfm->node = node;
>
> err = frontend->init_tfm(tfm);
> if (err)
> @@ -472,7 +474,7 @@ void *crypto_create_tfm(struct crypto_alg *alg,
> out:
> return mem;
> }
> -EXPORT_SYMBOL_GPL(crypto_create_tfm);
> +EXPORT_SYMBOL_GPL(crypto_create_tfm_node);
>
> struct crypto_alg *crypto_find_alg(const char *alg_name,
> const struct crypto_type *frontend,
> @@ -490,11 +492,13 @@ struct crypto_alg *crypto_find_alg(const char *alg_name,
> EXPORT_SYMBOL_GPL(crypto_find_alg);
>
> /*
> - * crypto_alloc_tfm - Locate algorithm and allocate transform
> + * crypto_alloc_tfm_node - Locate algorithm and allocate transform
> * @alg_name: Name of algorithm
> * @frontend: Frontend algorithm type
> * @type: Type of algorithm
> * @mask: Mask for type comparison
> + * @node: NUMA node in which users desire to put requests, if node is
> + * NUMA_NO_NODE, it means users have no special requirement.
> *
> * crypto_alloc_tfm() will first attempt to locate an already loaded
> * algorithm. If that fails and the kernel supports dynamically loadable
> @@ -509,8 +513,10 @@ EXPORT_SYMBOL_GPL(crypto_find_alg);
> *
> * In case of error the return value is an error pointer.
> */
> -void *crypto_alloc_tfm(const char *alg_name,
> - const struct crypto_type *frontend, u32 type, u32 mask)
> +
> +void *crypto_alloc_tfm_node(const char *alg_name,
> + const struct crypto_type *frontend, u32 type, u32 mask,
> + int node)
> {
> void *tfm;
> int err;
> @@ -524,7 +530,7 @@ void *crypto_alloc_tfm(const char *alg_name,
> goto err;
> }
>
> - tfm = crypto_create_tfm(alg, frontend);
> + tfm = crypto_create_tfm_node(alg, frontend, node);
> if (!IS_ERR(tfm))
> return tfm;
>
> @@ -542,7 +548,7 @@ void *crypto_alloc_tfm(const char *alg_name,
>
> return ERR_PTR(err);
> }
> -EXPORT_SYMBOL_GPL(crypto_alloc_tfm);
> +EXPORT_SYMBOL_GPL(crypto_alloc_tfm_node);
>
> /*
> * crypto_destroy_tfm - Free crypto transform
> diff --git a/crypto/internal.h b/crypto/internal.h
> index ff06a3bd1ca1..1b92a5a61852 100644
> --- a/crypto/internal.h
> +++ b/crypto/internal.h
> @@ -68,13 +68,28 @@ void crypto_remove_final(struct list_head *list);
> void crypto_shoot_alg(struct crypto_alg *alg);
> struct crypto_tfm *__crypto_alloc_tfm(struct crypto_alg *alg, u32 type,
> u32 mask);
> -void *crypto_create_tfm(struct crypto_alg *alg,
> - const struct crypto_type *frontend);
> +void *crypto_create_tfm_node(struct crypto_alg *alg,
> + const struct crypto_type *frontend, int node);
> +
> +static inline void *crypto_create_tfm(struct crypto_alg *alg,
> + const struct crypto_type *frontend)
> +{
> + return crypto_create_tfm_node(alg, frontend, NUMA_NO_NODE);
> +}
> +
> struct crypto_alg *crypto_find_alg(const char *alg_name,
> const struct crypto_type *frontend,
> u32 type, u32 mask);
> -void *crypto_alloc_tfm(const char *alg_name,
> - const struct crypto_type *frontend, u32 type, u32 mask);
> +
> +void *crypto_alloc_tfm_node(const char *alg_name,
> + const struct crypto_type *frontend, u32 type, u32 mask,
> + int node);
> +
> +static inline void *crypto_alloc_tfm(const char *alg_name,
> + const struct crypto_type *frontend, u32 type, u32 mask)
> +{
> + return crypto_alloc_tfm_node(alg_name, frontend, type, mask, NUMA_NO_NODE);
> +}
>
> int crypto_probing_notify(unsigned long val, void *v);
>
> diff --git a/include/crypto/acompress.h b/include/crypto/acompress.h
> index 2b4d2b06ccbd..b1a78687014a 100644
> --- a/include/crypto/acompress.h
> +++ b/include/crypto/acompress.h
> @@ -106,6 +106,13 @@ struct acomp_alg {
> */
> struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
> u32 mask);
> +/**
> + * crypto_alloc_acomp_node() -- allocate ACOMPRESS tfm handle with desired NUMA
> + * node
Given slightly relaxed view on 80 chars now in place, I'd put that on one line.
Also kernel-doc needs to be complete so though it's tedious you should document
the other parameters.
> + * @node: specifies the NUMA node the ZIP hardware belongs to
> + */
> +struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
> + u32 mask, int node);
>
> static inline struct crypto_tfm *crypto_acomp_tfm(struct crypto_acomp *tfm)
> {
> diff --git a/include/linux/crypto.h b/include/linux/crypto.h
> index 763863dbc079..c1a47ce4c09e 100644
> --- a/include/linux/crypto.h
> +++ b/include/linux/crypto.h
> @@ -593,8 +593,9 @@ int crypto_has_alg(const char *name, u32 type, u32 mask);
> */
>
> struct crypto_tfm {
> -
Stray change. Shouldn't be in this patch.
> u32 crt_flags;
> +
> + int node;
>
> void (*exit)(struct crypto_tfm *tfm);
>
Powered by blists - more mailing lists