[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20160307.140522.1910222129785231486.davem@davemloft.net>
Date: Mon, 07 Mar 2016 14:05:22 -0500 (EST)
From: David Miller <davem@...emloft.net>
To: sd@...asysnail.net
Cc: netdev@...r.kernel.org, hannes@...essinduktion.org, fw@...len.de,
pabeni@...hat.com, stephen@...workplumber.org
Subject: Re: [PATCH net-next 3/3] macsec: introduce IEEE 802.1AE driver
From: Sabrina Dubroca <sd@...asysnail.net>
Date: Mon, 7 Mar 2016 18:12:40 +0100
> diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig
> index f184fb5bd110..2a1ba62b7da2 100644
> --- a/drivers/net/Kconfig
> +++ b/drivers/net/Kconfig
> @@ -193,6 +193,13 @@ config GENEVE
> To compile this driver as a module, choose M here: the module
> will be called geneve.
>
> +config MACSEC
> + tristate "IEEE 802.1AE MAC-level encryption (MACsec)"
> + select CRYPTO_AES
> + select CRYPTO_GCM
> + ---help---
> + MACsec is an encryption standard for Ethernet.
> +
> config NETCONSOLE
> tristate "Network console logging support"
> ---help---
> diff --git a/drivers/net/Makefile b/drivers/net/Makefile
> index 900b0c5320bb..1aa7cb845663 100644
> --- a/drivers/net/Makefile
> +++ b/drivers/net/Makefile
> @@ -10,6 +10,7 @@ obj-$(CONFIG_IPVLAN) += ipvlan/
> obj-$(CONFIG_DUMMY) += dummy.o
> obj-$(CONFIG_EQUALIZER) += eql.o
> obj-$(CONFIG_IFB) += ifb.o
> +obj-$(CONFIG_MACSEC) += macsec.o
> obj-$(CONFIG_MACVLAN) += macvlan.o
> obj-$(CONFIG_MACVTAP) += macvtap.o
> obj-$(CONFIG_MII) += mii.o
> diff --git a/drivers/net/macsec.c b/drivers/net/macsec.c
> new file mode 100644
> index 000000000000..af79f59d4dd7
> --- /dev/null
> +++ b/drivers/net/macsec.c
> @@ -0,0 +1,3037 @@
> +/*
> + * drivers/net/macsec.c - MACsec device
> + *
> + * Copyright (c) 2015 Sabrina Dubroca <sd@...asysnail.net>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + */
> +
> +#include <linux/types.h>
> +#include <linux/skbuff.h>
> +#include <linux/socket.h>
> +#include <linux/module.h>
> +#include <crypto/aead.h>
> +#include <linux/etherdevice.h>
> +#include <linux/rtnetlink.h>
> +#include <net/genetlink.h>
> +#include <net/sock.h>
> +
> +#include <uapi/linux/if_macsec.h>
> +
> +typedef u64 __bitwise sci_t;
> +
> +#define MACSEC_SCI_LEN 8
> +
> +/* SecTAG length = macsec_eth_header without the optional SCI */
> +#define MACSEC_TAG_LEN 6
> +
> +struct macsec_eth_header {
> + struct ethhdr eth;
> + /* SecTAG */
> + u8 tci_an;
> +#if defined(__LITTLE_ENDIAN_BITFIELD)
> + u8 short_length:6,
> + unused:2;
> +#elif defined(__BIG_ENDIAN_BITFIELD)
> + u8 unused:2,
> + short_length:6;
> +#else
> +#error "Please fix <asm/byteorder.h>"
> +#endif
> + __be32 packet_number;
> + u8 secure_channel_id[8]; /* optional */
> +} __packed;
> +
> +#define MACSEC_TCI_VERSION 0x80
> +#define MACSEC_TCI_ES 0x40 /* end station */
> +#define MACSEC_TCI_SC 0x20 /* SCI present */
> +#define MACSEC_TCI_SCB 0x10 /* epon */
> +#define MACSEC_TCI_E 0x08 /* encryption */
> +#define MACSEC_TCI_C 0x04 /* changed text */
> +#define MACSEC_AN_MASK 0x03 /* association number */
> +#define MACSEC_TCI_CONFID (MACSEC_TCI_E | MACSEC_TCI_C)
> +
> +#define MACSEC_SHORTLEN_THR 48
> +
> +#define GCM_AES_IV_LEN 12
> +#define DEFAULT_ICV_LEN 16
> +
> +#define for_each_rxsc(secy, sc) \
> + for (sc = rcu_dereference_bh(secy->rx_sc); \
> + sc; \
> + sc = rcu_dereference_bh(sc->next))
> +#define for_each_rxsc_rtnl(secy, sc) \
> + for (sc = rtnl_dereference(secy->rx_sc); \
> + sc; \
> + sc = rtnl_dereference(sc->next))
> +
> +struct gcm_iv {
> + union {
> + u8 secure_channel_id[8];
> + sci_t sci;
> + };
> + __be32 pn;
> +};
> +
> +/**
> + * struct macsec_key - SA key
> + * @id user-provided key identifier
> + * @tfm crypto struct, key storage
> + */
> +struct macsec_key {
> + u64 id;
> + struct crypto_aead *tfm;
> +};
> +
> +/**
> + * struct macsec_rx_sa - receive secure association
> + * @active
> + * @next_pn packet number expected for the next packet
> + * @lock protects next_pn manipulations
> + * @key key structure
> + * @stats per-SA stats
> + */
> +struct macsec_rx_sa {
> + bool active;
> + u32 next_pn;
> + spinlock_t lock;
> + struct macsec_key key;
> + struct macsec_rx_sa_stats __percpu *stats;
> + struct macsec_rx_sc *sc;
> + atomic_t refcnt;
> + struct rcu_head rcu;
> +};
> +
> +struct pcpu_rx_sc_stats {
> + struct macsec_rx_sc_stats stats;
> + struct u64_stats_sync syncp;
> +};
> +
> +/**
> + * struct macsec_rx_sc - receive secure channel
> + * @sci secure channel identifier for this SC
> + * @active channel is active
> + * @sa array of secure associations
> + * @stats per-SC stats
> + */
> +struct macsec_rx_sc {
> + struct macsec_rx_sc __rcu *next;
> + sci_t sci;
> + bool active;
> + struct macsec_rx_sa __rcu *sa[4];
> + struct pcpu_rx_sc_stats __percpu *stats;
> + atomic_t refcnt;
> + struct rcu_head rcu_head;
> +};
> +
> +/**
> + * struct macsec_tx_sa - transmit secure association
> + * @active
> + * @next_pn packet number to use for the next packet
> + * @lock protects next_pn manipulations
> + * @key key structure
> + * @stats per-SA stats
> + */
> +struct macsec_tx_sa {
> + bool active;
> + u32 next_pn;
> + spinlock_t lock;
> + struct macsec_key key;
> + struct macsec_tx_sa_stats __percpu *stats;
> + atomic_t refcnt;
> + struct rcu_head rcu;
> +};
> +
> +struct pcpu_tx_sc_stats {
> + struct macsec_tx_sc_stats stats;
> + struct u64_stats_sync syncp;
> +};
> +
> +/**
> + * struct macsec_tx_sc - transmit secure channel
> + * @active
> + * @encoding_sa association number of the SA currently in use
> + * @encrypt encrypt packets on transmit, or authenticate only
> + * @send_sci always include the SCI in the SecTAG
> + * @end_station
> + * @scb single copy broadcast flag
> + * @sa array of secure associations
> + * @stats stats for this TXSC
> + */
> +struct macsec_tx_sc {
> + bool active;
> + u8 encoding_sa;
> + bool encrypt;
> + bool send_sci;
> + bool end_station;
> + bool scb;
> + struct macsec_tx_sa __rcu *sa[4];
> + struct pcpu_tx_sc_stats __percpu *stats;
> +};
> +
> +#define MACSEC_VALIDATE_DEFAULT MACSEC_VALIDATE_STRICT
> +
> +/**
> + * struct macsec_secy - MACsec Security Entity
> + * @netdev netdevice for this SecY
> + * @n_rx_sc number of receive secure channels configured on this SecY
> + * @sci secure channel identifier used for tx
> + * @key_len length of keys used by the cipher suite
> + * @icv_len length of ICV used by the cipher suite
> + * @validate_frames validation mode
> + * @operational MAC_Operational flag
> + * @protect_frames enable protection for this SecY
> + * @replay_protect enable packet number checks on receive
> + * @replay_window size of the replay window
> + * @tx_sc transmit secure channel
> + * @rx_sc linked list of receive secure channels
> + */
> +struct macsec_secy {
> + struct net_device *netdev;
> + unsigned int n_rx_sc;
> + sci_t sci;
> + u16 key_len;
> + u16 icv_len;
> + enum validation_type validate_frames;
> + bool operational;
> + bool protect_frames;
> + bool replay_protect;
> + u32 replay_window;
> + struct macsec_tx_sc tx_sc;
> + struct macsec_rx_sc __rcu *rx_sc;
> +};
> +
> +struct pcpu_secy_stats {
> + struct macsec_dev_stats stats;
> + struct u64_stats_sync syncp;
> +};
> +
> +/**
> + * struct macsec_dev - private data
> + * @secy SecY config
> + * @real_dev pointer to underlying netdevice
> + * @stats MACsec device stats
> + * @secys linked list of SecY's on the underlying device
> + */
> +struct macsec_dev {
> + struct macsec_secy secy;
> + struct net_device *real_dev;
> + struct pcpu_secy_stats __percpu *stats;
> + struct list_head secys;
> +};
> +
> +/**
> + * struct macsec_rxh_data - rx_handler private argument
> + * @secys linked list of SecY's on this underlying device
> + */
> +struct macsec_rxh_data {
> + struct list_head secys;
> +};
> +
> +static struct macsec_dev *macsec_priv(const struct net_device *dev)
> +{
> + return (struct macsec_dev *)netdev_priv(dev);
> +}
> +
> +static struct macsec_rxh_data *macsec_data_rcu(const struct net_device *dev)
> +{
> + return rcu_dereference_bh(dev->rx_handler_data);
> +}
> +
> +static struct macsec_rxh_data *macsec_data_rtnl(const struct net_device *dev)
> +{
> + return rtnl_dereference(dev->rx_handler_data);
> +}
> +
> +struct macsec_cb {
> + struct aead_request *req;
> + union {
> + struct macsec_tx_sa *tx_sa;
> + struct macsec_rx_sa *rx_sa;
> + };
> + u8 assoc_num;
> + bool valid;
> + bool has_sci;
> +};
> +
> +static struct macsec_rx_sa *macsec_rxsa_get(struct macsec_rx_sa __rcu *ptr)
> +{
> + struct macsec_rx_sa *sa = rcu_dereference_bh(ptr);
> +
> + if (!sa || !sa->active)
> + return NULL;
> +
> + if (!atomic_inc_not_zero(&sa->refcnt))
> + return NULL;
> +
> + return sa;
> +}
> +
> +static void free_rx_sc_rcu(struct rcu_head *head)
> +{
> + struct macsec_rx_sc *rx_sc = container_of(head, struct macsec_rx_sc, rcu_head);
> +
> + free_percpu(rx_sc->stats);
> + kfree(rx_sc);
> +}
> +
> +static struct macsec_rx_sc *macsec_rxsc_get(struct macsec_rx_sc *sc)
> +{
> + return atomic_inc_not_zero(&sc->refcnt) ? sc : NULL;
> +}
> +
> +static void macsec_rxsc_put(struct macsec_rx_sc *sc)
> +{
> + if (atomic_dec_and_test(&sc->refcnt))
> + call_rcu(&sc->rcu_head, free_rx_sc_rcu);
> +}
> +
> +static void free_rxsa(struct rcu_head *head)
> +{
> + struct macsec_rx_sa *sa = container_of(head, struct macsec_rx_sa, rcu);
> +
> + crypto_free_aead(sa->key.tfm);
> + free_percpu(sa->stats);
> + macsec_rxsc_put(sa->sc);
> + kfree(sa);
> +}
> +
> +static void macsec_rxsa_put(struct macsec_rx_sa *sa)
> +{
> + if (atomic_dec_and_test(&sa->refcnt))
> + call_rcu(&sa->rcu, free_rxsa);
> +}
> +
> +static struct macsec_tx_sa *macsec_txsa_get(struct macsec_tx_sa __rcu *ptr)
> +{
> + struct macsec_tx_sa *sa = rcu_dereference_bh(ptr);
> +
> + if (!sa || !sa->active)
> + return NULL;
> +
> + if (!atomic_inc_not_zero(&sa->refcnt))
> + return NULL;
> +
> + return sa;
> +}
> +
> +static void free_txsa(struct rcu_head *head)
> +{
> + struct macsec_tx_sa *sa = container_of(head, struct macsec_tx_sa, rcu);
> +
> + crypto_free_aead(sa->key.tfm);
> + free_percpu(sa->stats);
> + kfree(sa);
> +}
> +
> +static void macsec_txsa_put(struct macsec_tx_sa *sa)
> +{
> + if (atomic_dec_and_test(&sa->refcnt))
> + call_rcu(&sa->rcu, free_txsa);
> +}
> +
> +static struct macsec_cb *macsec_skb_cb(struct sk_buff *skb)
> +{
> + BUILD_BUG_ON(sizeof(struct macsec_cb) > sizeof(skb->cb));
> + return (struct macsec_cb *)skb->cb;
> +}
> +
> +#define MACSEC_PORT_ES (htons(0x0001))
> +#define MACSEC_PORT_SCB (0x0000)
> +#define MACSEC_UNDEF_SCI ((__force sci_t)0xffffffffffffffffULL)
> +
> +#define DEFAULT_SAK_LEN 16
> +#define DEFAULT_SEND_SCI true
> +#define DEFAULT_ENCRYPT false
> +#define DEFAULT_ENCODING_SA 0
> +
> +static sci_t make_sci(u8 *addr, __be16 port)
> +{
> + sci_t sci;
> +
> + memcpy(&sci, addr, ETH_ALEN);
> + memcpy(((char *)&sci) + ETH_ALEN, &port, sizeof(port));
> +
> + return sci;
> +}
> +
> +static sci_t macsec_frame_sci(struct macsec_eth_header *hdr, bool sci_present)
> +{
> + sci_t sci;
> +
> + if (sci_present)
> + memcpy(&sci, hdr->secure_channel_id,
> + sizeof(hdr->secure_channel_id));
> + else
> + sci = make_sci(hdr->eth.h_source, MACSEC_PORT_ES);
> +
> + return sci;
> +}
> +
> +static unsigned int macsec_sectag_len(bool sci_present)
> +{
> + return MACSEC_TAG_LEN + (sci_present ? MACSEC_SCI_LEN : 0);
> +}
> +
> +static unsigned int macsec_hdr_len(bool sci_present)
> +{
> + return macsec_sectag_len(sci_present) + ETH_HLEN;
> +}
> +
> +static unsigned int macsec_extra_len(bool sci_present)
> +{
> + return macsec_sectag_len(sci_present) + sizeof(__be16);
> +}
> +
> +/* Fill SecTAG according to IEEE 802.1AE-2006 10.5.3 */
> +static void macsec_fill_sectag(struct macsec_eth_header *h,
> + const struct macsec_secy *secy, u32 pn)
> +{
> + const struct macsec_tx_sc *tx_sc = &secy->tx_sc;
> +
> + memset(&h->tci_an, 0, macsec_sectag_len(tx_sc->send_sci));
> + h->eth.h_proto = htons(ETH_P_MACSEC);
> +
> + if (tx_sc->send_sci ||
> + (secy->n_rx_sc > 1 && !tx_sc->end_station && !tx_sc->scb)) {
> + h->tci_an |= MACSEC_TCI_SC;
> + memcpy(&h->secure_channel_id, &secy->sci,
> + sizeof(h->secure_channel_id));
> + } else {
> + if (tx_sc->end_station)
> + h->tci_an |= MACSEC_TCI_ES;
> + if (tx_sc->scb)
> + h->tci_an |= MACSEC_TCI_SCB;
> + }
> +
> + h->packet_number = htonl(pn);
> +
> + /* with GCM, C/E clear for !encrypt, both set for encrypt */
> + if (tx_sc->encrypt)
> + h->tci_an |= MACSEC_TCI_CONFID;
> + else if (secy->icv_len != DEFAULT_ICV_LEN)
> + h->tci_an |= MACSEC_TCI_C;
> +
> + h->tci_an |= tx_sc->encoding_sa;
> +}
> +
> +static void macsec_set_shortlen(struct macsec_eth_header *h, size_t data_len)
> +{
> + if (data_len < MACSEC_SHORTLEN_THR)
> + h->short_length = data_len;
> +}
> +
> +/* minimum secure data length deemed "not short", see IEEE 802.1AE-2006 9.7 */
> +#define MIN_NON_SHORT_LEN 48
> +
> +/* validate MACsec packet according to IEEE 802.1AE-2006 9.12 */
> +static bool macsec_validate_skb(struct sk_buff *skb, u16 icv_len)
> +{
> + struct macsec_eth_header *h = (struct macsec_eth_header *)skb->data;
> + int len = skb->len - 2 * ETH_ALEN;
> + int extra_len = macsec_extra_len(!!(h->tci_an & MACSEC_TCI_SC)) + icv_len;
> +
> + /* a) It comprises at least 17 octets */
> + if (skb->len <= 16)
> + return false;
> +
> + /* b) MACsec EtherType: already checked */
> +
> + /* c) V bit is clear */
> + if (h->tci_an & MACSEC_TCI_VERSION)
> + return false;
> +
> + /* d) ES or SCB => !SC */
> + if ((h->tci_an & MACSEC_TCI_ES || h->tci_an & MACSEC_TCI_SCB) &&
> + (h->tci_an & MACSEC_TCI_SC))
> + return false;
> +
> + /* e) Bits 7 and 8 of octet 4 of the SecTAG are clear */
> + if (h->unused)
> + return false;
> +
> + /* rx.pn != 0 (figure 10-5) */
> + if (!h->packet_number)
> + return false;
> +
> + /* length check, f) g) h) i) */
> + if (h->short_length)
> + return len == extra_len + h->short_length;
> + return len >= extra_len + MIN_NON_SHORT_LEN;
> +}
> +
> +#define MACSEC_NEEDED_HEADROOM (macsec_extra_len(true))
> +#define MACSEC_NEEDED_TAILROOM MACSEC_MAX_ICV_LEN
> +
> +static void macsec_fill_iv(unsigned char *iv, sci_t sci, u32 pn)
> +{
> + struct gcm_iv *gcm_iv = (struct gcm_iv *)iv;
> +
> + gcm_iv->sci = sci;
> + gcm_iv->pn = htonl(pn);
> +}
> +
> +static struct macsec_eth_header *macsec_ethhdr(struct sk_buff *skb)
> +{
> + return (struct macsec_eth_header *)skb_mac_header(skb);
> +}
> +
> +static u32 tx_sa_update_pn(struct macsec_tx_sa *tx_sa, struct macsec_secy *secy)
> +{
> + u32 pn;
> +
> + spin_lock_bh(&tx_sa->lock);
> + pn = tx_sa->next_pn;
> +
> + tx_sa->next_pn++;
> + if (tx_sa->next_pn == 0) {
> + pr_debug("PN wrapped, transitionning to !oper\n");
> + tx_sa->active = false;
> + if (secy->protect_frames)
> + secy->operational = false;
> + }
> + spin_unlock_bh(&tx_sa->lock);
> +
> + return pn;
> +}
> +
> +static void macsec_encrypt_finish(struct sk_buff *skb, struct net_device *dev)
> +{
> + struct macsec_dev *macsec = netdev_priv(dev);
> +
> + skb->dev = macsec->real_dev;
> + skb_reset_mac_header(skb);
> + skb->protocol = eth_hdr(skb)->h_proto;
> +}
> +
> +static void macsec_count_tx(struct sk_buff *skb, struct macsec_tx_sc *tx_sc,
> + struct macsec_tx_sa *tx_sa)
> +{
> + struct pcpu_tx_sc_stats *txsc_stats = this_cpu_ptr(tx_sc->stats);
> +
> + u64_stats_update_begin(&txsc_stats->syncp);
> + if (tx_sc->encrypt) {
> + txsc_stats->stats.OutOctetsEncrypted += skb->len;
> + txsc_stats->stats.OutPktsEncrypted++;
> + this_cpu_inc(tx_sa->stats->OutPktsEncrypted);
> + } else {
> + txsc_stats->stats.OutOctetsProtected += skb->len;
> + txsc_stats->stats.OutPktsProtected++;
> + this_cpu_inc(tx_sa->stats->OutPktsProtected);
> + }
> + u64_stats_update_end(&txsc_stats->syncp);
> +}
> +
> +static void count_tx(struct net_device *dev, int ret, int len)
> +{
> + if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) {
> + struct pcpu_sw_netstats *stats = this_cpu_ptr(dev->tstats);
> +
> + u64_stats_update_begin(&stats->syncp);
> + stats->tx_packets++;
> + stats->tx_bytes += len;
> + u64_stats_update_end(&stats->syncp);
> + } else {
> + dev->stats.tx_dropped++;
> + }
> +}
> +
> +static void macsec_encrypt_done(struct crypto_async_request *base, int err)
> +{
> + struct sk_buff *skb = base->data;
> + struct net_device *dev = skb->dev;
> + struct macsec_dev *macsec = macsec_priv(dev);
> + struct macsec_tx_sa *sa = macsec_skb_cb(skb)->tx_sa;
> + int len, ret;
> +
> + aead_request_free(macsec_skb_cb(skb)->req);
> +
> + rcu_read_lock_bh();
> + macsec_encrypt_finish(skb, dev);
> + macsec_count_tx(skb, &macsec->secy.tx_sc, macsec_skb_cb(skb)->tx_sa);
> + len = skb->len;
> + ret = dev_queue_xmit(skb);
> + count_tx(dev, ret, len);
> + rcu_read_unlock_bh();
> +
> + macsec_txsa_put(sa);
> + dev_put(dev);
> +}
> +
> +static struct sk_buff *macsec_encrypt(struct sk_buff *skb,
> + struct net_device *dev)
> +{
> + int ret;
> + struct scatterlist sg[MAX_SKB_FRAGS + 1];
> + unsigned char iv[GCM_AES_IV_LEN];
> + struct ethhdr *eth;
> + struct macsec_eth_header *hh;
> + size_t unprotected_len;
> + struct aead_request *req;
> + struct macsec_secy *secy;
> + struct macsec_tx_sc *tx_sc;
> + struct macsec_tx_sa *tx_sa;
> + struct macsec_dev *macsec = macsec_priv(dev);
> + u32 pn;
> +
> + secy = &macsec->secy;
> + tx_sc = &secy->tx_sc;
> +
> + /* 10.5.1 TX SA assignment */
> + tx_sa = macsec_txsa_get(tx_sc->sa[tx_sc->encoding_sa]);
> + if (!tx_sa) {
> + secy->operational = false;
> + kfree_skb(skb);
> + return ERR_PTR(-EINVAL);
> + }
> +
> + if (unlikely(skb_headroom(skb) < MACSEC_NEEDED_HEADROOM ||
> + skb_tailroom(skb) < MACSEC_NEEDED_TAILROOM)) {
> + struct sk_buff *nskb = skb_copy_expand(skb,
> + MACSEC_NEEDED_HEADROOM,
> + MACSEC_NEEDED_TAILROOM,
> + GFP_ATOMIC);
> + if (likely(nskb)) {
> + consume_skb(skb);
> + skb = nskb;
> + } else {
> + macsec_txsa_put(tx_sa);
> + kfree_skb(skb);
> + return ERR_PTR(-ENOMEM);
> + }
> + } else {
> + skb = skb_unshare(skb, GFP_ATOMIC);
> + if (!skb) {
> + macsec_txsa_put(tx_sa);
> + return ERR_PTR(-ENOMEM);
> + }
> + }
> +
> + unprotected_len = skb->len;
> + eth = eth_hdr(skb);
> + hh = (struct macsec_eth_header *)skb_push(skb, macsec_extra_len(tx_sc->send_sci));
> + memmove(hh, eth, 2 * ETH_ALEN);
> +
> + pn = tx_sa_update_pn(tx_sa, secy);
> + if (pn == 0) {
> + macsec_txsa_put(tx_sa);
> + kfree_skb(skb);
> + return ERR_PTR(-ENOLINK);
> + }
> + macsec_fill_sectag(hh, secy, pn);
> + macsec_set_shortlen(hh, unprotected_len - 2 * ETH_ALEN);
> +
> + macsec_fill_iv(iv, secy->sci, pn);
> +
> + skb_put(skb, secy->icv_len);
> +
> + if (skb->len - ETH_HLEN > macsec_priv(dev)->real_dev->mtu) {
> + struct pcpu_secy_stats *secy_stats = this_cpu_ptr(macsec->stats);
> +
> + u64_stats_update_begin(&secy_stats->syncp);
> + secy_stats->stats.OutPktsTooLong++;
> + u64_stats_update_end(&secy_stats->syncp);
> +
> + macsec_txsa_put(tx_sa);
> + kfree_skb(skb);
> + return ERR_PTR(-EINVAL);
> + }
> +
> + req = aead_request_alloc(tx_sa->key.tfm, GFP_ATOMIC);
> + if (!req) {
> + macsec_txsa_put(tx_sa);
> + kfree_skb(skb);
> + return ERR_PTR(-ENOMEM);
> + }
> +
> + sg_init_table(sg, MAX_SKB_FRAGS + 1);
> + skb_to_sgvec(skb, sg, 0, skb->len);
> +
> + if (tx_sc->encrypt) {
> + int len = skb->len - macsec_hdr_len(tx_sc->send_sci) -
> + secy->icv_len;
> + aead_request_set_crypt(req, sg, sg, len, iv);
> + aead_request_set_ad(req, macsec_hdr_len(tx_sc->send_sci));
> + } else {
> + aead_request_set_crypt(req, sg, sg, 0, iv);
> + aead_request_set_ad(req, skb->len - secy->icv_len);
> + }
> +
> + macsec_skb_cb(skb)->req = req;
> + macsec_skb_cb(skb)->tx_sa = tx_sa;
> + aead_request_set_callback(req, 0, macsec_encrypt_done, skb);
> +
> + dev_hold(skb->dev);
> + ret = crypto_aead_encrypt(req);
> + if (ret == -EINPROGRESS) {
> + return ERR_PTR(ret);
> + } else if (ret != 0) {
> + dev_put(skb->dev);
> + kfree_skb(skb);
> + aead_request_free(req);
> + macsec_txsa_put(tx_sa);
> + return ERR_PTR(-EINVAL);
> + }
> +
> + dev_put(skb->dev);
> + aead_request_free(req);
> + macsec_txsa_put(tx_sa);
> +
> + return skb;
> +}
> +
> +static bool macsec_post_decrypt(struct sk_buff *skb, struct macsec_secy *secy, u32 pn)
> +{
> + struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa;
> + struct pcpu_rx_sc_stats *rxsc_stats = this_cpu_ptr(rx_sa->sc->stats);
> + struct macsec_eth_header *hdr = macsec_ethhdr(skb);
> + u32 lowest_pn = 0;
> +
> + spin_lock(&rx_sa->lock);
> + if (rx_sa->next_pn >= secy->replay_window)
> + lowest_pn = rx_sa->next_pn - secy->replay_window;
> +
> + /* Now perform replay protection check again
> + * (see IEEE 802.1AE-2006 figure 10-5)
> + */
> + if (secy->replay_protect && pn < lowest_pn) {
> + spin_unlock(&rx_sa->lock);
> + u64_stats_update_begin(&rxsc_stats->syncp);
> + rxsc_stats->stats.InPktsLate++;
> + u64_stats_update_end(&rxsc_stats->syncp);
> + return false;
> + }
> +
> + if (secy->validate_frames != MACSEC_VALIDATE_DISABLED) {
> + u64_stats_update_begin(&rxsc_stats->syncp);
> + if (hdr->tci_an & MACSEC_TCI_E)
> + rxsc_stats->stats.InOctetsDecrypted += skb->len;
> + else
> + rxsc_stats->stats.InOctetsValidated += skb->len;
> + u64_stats_update_end(&rxsc_stats->syncp);
> + }
> +
> + if (!macsec_skb_cb(skb)->valid) {
> + spin_unlock(&rx_sa->lock);
> +
> + /* 10.6.5 */
> + if (hdr->tci_an & MACSEC_TCI_C ||
> + secy->validate_frames == MACSEC_VALIDATE_STRICT) {
> + u64_stats_update_begin(&rxsc_stats->syncp);
> + rxsc_stats->stats.InPktsNotValid++;
> + u64_stats_update_end(&rxsc_stats->syncp);
> + return false;
> + }
> +
> + u64_stats_update_begin(&rxsc_stats->syncp);
> + if (secy->validate_frames == MACSEC_VALIDATE_CHECK) {
> + rxsc_stats->stats.InPktsInvalid++;
> + this_cpu_inc(rx_sa->stats->InPktsInvalid);
> + } else if (pn < lowest_pn) {
> + rxsc_stats->stats.InPktsDelayed++;
> + } else {
> + rxsc_stats->stats.InPktsUnchecked++;
> + }
> + u64_stats_update_end(&rxsc_stats->syncp);
> + } else {
> + u64_stats_update_begin(&rxsc_stats->syncp);
> + if (pn < lowest_pn) {
> + rxsc_stats->stats.InPktsDelayed++;
> + } else {
> + rxsc_stats->stats.InPktsOK++;
> + this_cpu_inc(rx_sa->stats->InPktsOK);
> + }
> + u64_stats_update_end(&rxsc_stats->syncp);
> +
> + if (pn >= rx_sa->next_pn)
> + rx_sa->next_pn = pn + 1;
> + spin_unlock(&rx_sa->lock);
> + }
> +
> + return true;
> +}
> +
> +static void macsec_reset_skb(struct sk_buff *skb, struct net_device *dev)
> +{
> + skb->pkt_type = PACKET_HOST;
> + skb->protocol = eth_type_trans(skb, dev);
> +
> + skb_reset_network_header(skb);
> + if (!skb_transport_header_was_set(skb))
> + skb_reset_transport_header(skb);
> + skb_reset_mac_len(skb);
> +}
> +
> +static void macsec_finalize_skb(struct sk_buff *skb, u8 icv_len, u8 hdr_len)
> +{
> + memmove(skb->data + hdr_len, skb->data, 2 * ETH_ALEN);
> + skb_pull(skb, hdr_len);
> + pskb_trim_unique(skb, skb->len - icv_len);
> +}
> +
> +static void count_rx(struct net_device *dev, int len)
> +{
> + struct pcpu_sw_netstats *stats = this_cpu_ptr(dev->tstats);
> +
> + u64_stats_update_begin(&stats->syncp);
> + stats->rx_packets++;
> + stats->rx_bytes += len;
> + u64_stats_update_end(&stats->syncp);
> +}
> +
> +static void macsec_decrypt_done(struct crypto_async_request *base, int err)
> +{
> + struct sk_buff *skb = base->data;
> + struct net_device *dev = skb->dev;
> + struct macsec_dev *macsec = macsec_priv(dev);
> + struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa;
> + int len, ret;
> + u32 pn;
> +
> + aead_request_free(macsec_skb_cb(skb)->req);
> +
> + rcu_read_lock_bh();
> + pn = ntohl(macsec_ethhdr(skb)->packet_number);
> + if (!macsec_post_decrypt(skb, &macsec->secy, pn)) {
> + rcu_read_unlock_bh();
> + kfree_skb(skb);
> + goto out;
> + }
> +
> + macsec_finalize_skb(skb, macsec->secy.icv_len,
> + macsec_extra_len(macsec_skb_cb(skb)->has_sci));
> + macsec_reset_skb(skb, macsec->secy.netdev);
> +
> + len = skb->len;
> + ret = netif_rx(skb);
> + if (ret == NET_RX_SUCCESS)
> + count_rx(dev, len);
> + else
> + macsec->secy.netdev->stats.rx_dropped++;
> +
> + rcu_read_unlock_bh();
> +
> +out:
> + macsec_rxsa_put(rx_sa);
> + dev_put(dev);
> + return;
> +}
> +
> +static struct sk_buff *macsec_decrypt(struct sk_buff *skb,
> + struct net_device *dev,
> + struct macsec_rx_sa *rx_sa,
> + sci_t sci,
> + struct macsec_secy *secy)
> +{
> + int ret;
> + struct scatterlist sg[MAX_SKB_FRAGS + 1];
> + unsigned char iv[GCM_AES_IV_LEN];
> + struct aead_request *req;
> + struct macsec_eth_header *hdr;
> + u16 icv_len = secy->icv_len;
> +
> + macsec_skb_cb(skb)->valid = 0;
Please use true/false for boolean variables.
> + macsec_skb_cb(skb)->valid = 1;
Likewise.
> +static void handle_not_macsec(struct sk_buff *skb)
> +{
> + struct macsec_rxh_data *rxd;
> + struct macsec_dev *macsec;
> +
> + rcu_read_lock_bh();
"bh" should be implicit in this receive code path, so plain rcu_read_lock() should
be sufficient.
> +static rx_handler_result_t macsec_handle_frame(struct sk_buff **pskb)
> +{
...
> + rcu_read_lock_bh();
Likewise.
Otherwise looks really good to me.
Powered by blists - more mailing lists