From: Sabrina Dubroca <s...@queasysnail.net> Date: Mon, 7 Mar 2016 18:12:40 +0100
> diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig > index f184fb5bd110..2a1ba62b7da2 100644 > --- a/drivers/net/Kconfig > +++ b/drivers/net/Kconfig > @@ -193,6 +193,13 @@ config GENEVE > To compile this driver as a module, choose M here: the module > will be called geneve. > > +config MACSEC > + tristate "IEEE 802.1AE MAC-level encryption (MACsec)" > + select CRYPTO_AES > + select CRYPTO_GCM > + ---help--- > + MACsec is an encryption standard for Ethernet. > + > config NETCONSOLE > tristate "Network console logging support" > ---help--- > diff --git a/drivers/net/Makefile b/drivers/net/Makefile > index 900b0c5320bb..1aa7cb845663 100644 > --- a/drivers/net/Makefile > +++ b/drivers/net/Makefile > @@ -10,6 +10,7 @@ obj-$(CONFIG_IPVLAN) += ipvlan/ > obj-$(CONFIG_DUMMY) += dummy.o > obj-$(CONFIG_EQUALIZER) += eql.o > obj-$(CONFIG_IFB) += ifb.o > +obj-$(CONFIG_MACSEC) += macsec.o > obj-$(CONFIG_MACVLAN) += macvlan.o > obj-$(CONFIG_MACVTAP) += macvtap.o > obj-$(CONFIG_MII) += mii.o > diff --git a/drivers/net/macsec.c b/drivers/net/macsec.c > new file mode 100644 > index 000000000000..af79f59d4dd7 > --- /dev/null > +++ b/drivers/net/macsec.c > @@ -0,0 +1,3037 @@ > +/* > + * drivers/net/macsec.c - MACsec device > + * > + * Copyright (c) 2015 Sabrina Dubroca <s...@queasysnail.net> > + * > + * This program is free software; you can redistribute it and/or modify > + * it under the terms of the GNU General Public License as published by > + * the Free Software Foundation; either version 2 of the License, or > + * (at your option) any later version. > + */ > + > +#include <linux/types.h> > +#include <linux/skbuff.h> > +#include <linux/socket.h> > +#include <linux/module.h> > +#include <crypto/aead.h> > +#include <linux/etherdevice.h> > +#include <linux/rtnetlink.h> > +#include <net/genetlink.h> > +#include <net/sock.h> > + > +#include <uapi/linux/if_macsec.h> > + > +typedef u64 __bitwise sci_t; > + > +#define MACSEC_SCI_LEN 8 > + > +/* SecTAG length = macsec_eth_header without the optional SCI */ > +#define MACSEC_TAG_LEN 6 > + > +struct macsec_eth_header { > + struct ethhdr eth; > + /* SecTAG */ > + u8 tci_an; > +#if defined(__LITTLE_ENDIAN_BITFIELD) > + u8 short_length:6, > + unused:2; > +#elif defined(__BIG_ENDIAN_BITFIELD) > + u8 unused:2, > + short_length:6; > +#else > +#error "Please fix <asm/byteorder.h>" > +#endif > + __be32 packet_number; > + u8 secure_channel_id[8]; /* optional */ > +} __packed; > + > +#define MACSEC_TCI_VERSION 0x80 > +#define MACSEC_TCI_ES 0x40 /* end station */ > +#define MACSEC_TCI_SC 0x20 /* SCI present */ > +#define MACSEC_TCI_SCB 0x10 /* epon */ > +#define MACSEC_TCI_E 0x08 /* encryption */ > +#define MACSEC_TCI_C 0x04 /* changed text */ > +#define MACSEC_AN_MASK 0x03 /* association number */ > +#define MACSEC_TCI_CONFID (MACSEC_TCI_E | MACSEC_TCI_C) > + > +#define MACSEC_SHORTLEN_THR 48 > + > +#define GCM_AES_IV_LEN 12 > +#define DEFAULT_ICV_LEN 16 > + > +#define for_each_rxsc(secy, sc) \ > + for (sc = rcu_dereference_bh(secy->rx_sc); \ > + sc; \ > + sc = rcu_dereference_bh(sc->next)) > +#define for_each_rxsc_rtnl(secy, sc) \ > + for (sc = rtnl_dereference(secy->rx_sc); \ > + sc; \ > + sc = rtnl_dereference(sc->next)) > + > +struct gcm_iv { > + union { > + u8 secure_channel_id[8]; > + sci_t sci; > + }; > + __be32 pn; > +}; > + > +/** > + * struct macsec_key - SA key > + * @id user-provided key identifier > + * @tfm crypto struct, key storage > + */ > +struct macsec_key { > + u64 id; > + struct crypto_aead *tfm; > +}; > + > +/** > + * struct macsec_rx_sa - receive secure association > + * @active > + * @next_pn packet number expected for the next packet > + * @lock protects next_pn manipulations > + * @key key structure > + * @stats per-SA stats > + */ > +struct macsec_rx_sa { > + bool active; > + u32 next_pn; > + spinlock_t lock; > + struct macsec_key key; > + struct macsec_rx_sa_stats __percpu *stats; > + struct macsec_rx_sc *sc; > + atomic_t refcnt; > + struct rcu_head rcu; > +}; > + > +struct pcpu_rx_sc_stats { > + struct macsec_rx_sc_stats stats; > + struct u64_stats_sync syncp; > +}; > + > +/** > + * struct macsec_rx_sc - receive secure channel > + * @sci secure channel identifier for this SC > + * @active channel is active > + * @sa array of secure associations > + * @stats per-SC stats > + */ > +struct macsec_rx_sc { > + struct macsec_rx_sc __rcu *next; > + sci_t sci; > + bool active; > + struct macsec_rx_sa __rcu *sa[4]; > + struct pcpu_rx_sc_stats __percpu *stats; > + atomic_t refcnt; > + struct rcu_head rcu_head; > +}; > + > +/** > + * struct macsec_tx_sa - transmit secure association > + * @active > + * @next_pn packet number to use for the next packet > + * @lock protects next_pn manipulations > + * @key key structure > + * @stats per-SA stats > + */ > +struct macsec_tx_sa { > + bool active; > + u32 next_pn; > + spinlock_t lock; > + struct macsec_key key; > + struct macsec_tx_sa_stats __percpu *stats; > + atomic_t refcnt; > + struct rcu_head rcu; > +}; > + > +struct pcpu_tx_sc_stats { > + struct macsec_tx_sc_stats stats; > + struct u64_stats_sync syncp; > +}; > + > +/** > + * struct macsec_tx_sc - transmit secure channel > + * @active > + * @encoding_sa association number of the SA currently in use > + * @encrypt encrypt packets on transmit, or authenticate only > + * @send_sci always include the SCI in the SecTAG > + * @end_station > + * @scb single copy broadcast flag > + * @sa array of secure associations > + * @stats stats for this TXSC > + */ > +struct macsec_tx_sc { > + bool active; > + u8 encoding_sa; > + bool encrypt; > + bool send_sci; > + bool end_station; > + bool scb; > + struct macsec_tx_sa __rcu *sa[4]; > + struct pcpu_tx_sc_stats __percpu *stats; > +}; > + > +#define MACSEC_VALIDATE_DEFAULT MACSEC_VALIDATE_STRICT > + > +/** > + * struct macsec_secy - MACsec Security Entity > + * @netdev netdevice for this SecY > + * @n_rx_sc number of receive secure channels configured on this SecY > + * @sci secure channel identifier used for tx > + * @key_len length of keys used by the cipher suite > + * @icv_len length of ICV used by the cipher suite > + * @validate_frames validation mode > + * @operational MAC_Operational flag > + * @protect_frames enable protection for this SecY > + * @replay_protect enable packet number checks on receive > + * @replay_window size of the replay window > + * @tx_sc transmit secure channel > + * @rx_sc linked list of receive secure channels > + */ > +struct macsec_secy { > + struct net_device *netdev; > + unsigned int n_rx_sc; > + sci_t sci; > + u16 key_len; > + u16 icv_len; > + enum validation_type validate_frames; > + bool operational; > + bool protect_frames; > + bool replay_protect; > + u32 replay_window; > + struct macsec_tx_sc tx_sc; > + struct macsec_rx_sc __rcu *rx_sc; > +}; > + > +struct pcpu_secy_stats { > + struct macsec_dev_stats stats; > + struct u64_stats_sync syncp; > +}; > + > +/** > + * struct macsec_dev - private data > + * @secy SecY config > + * @real_dev pointer to underlying netdevice > + * @stats MACsec device stats > + * @secys linked list of SecY's on the underlying device > + */ > +struct macsec_dev { > + struct macsec_secy secy; > + struct net_device *real_dev; > + struct pcpu_secy_stats __percpu *stats; > + struct list_head secys; > +}; > + > +/** > + * struct macsec_rxh_data - rx_handler private argument > + * @secys linked list of SecY's on this underlying device > + */ > +struct macsec_rxh_data { > + struct list_head secys; > +}; > + > +static struct macsec_dev *macsec_priv(const struct net_device *dev) > +{ > + return (struct macsec_dev *)netdev_priv(dev); > +} > + > +static struct macsec_rxh_data *macsec_data_rcu(const struct net_device *dev) > +{ > + return rcu_dereference_bh(dev->rx_handler_data); > +} > + > +static struct macsec_rxh_data *macsec_data_rtnl(const struct net_device *dev) > +{ > + return rtnl_dereference(dev->rx_handler_data); > +} > + > +struct macsec_cb { > + struct aead_request *req; > + union { > + struct macsec_tx_sa *tx_sa; > + struct macsec_rx_sa *rx_sa; > + }; > + u8 assoc_num; > + bool valid; > + bool has_sci; > +}; > + > +static struct macsec_rx_sa *macsec_rxsa_get(struct macsec_rx_sa __rcu *ptr) > +{ > + struct macsec_rx_sa *sa = rcu_dereference_bh(ptr); > + > + if (!sa || !sa->active) > + return NULL; > + > + if (!atomic_inc_not_zero(&sa->refcnt)) > + return NULL; > + > + return sa; > +} > + > +static void free_rx_sc_rcu(struct rcu_head *head) > +{ > + struct macsec_rx_sc *rx_sc = container_of(head, struct macsec_rx_sc, > rcu_head); > + > + free_percpu(rx_sc->stats); > + kfree(rx_sc); > +} > + > +static struct macsec_rx_sc *macsec_rxsc_get(struct macsec_rx_sc *sc) > +{ > + return atomic_inc_not_zero(&sc->refcnt) ? sc : NULL; > +} > + > +static void macsec_rxsc_put(struct macsec_rx_sc *sc) > +{ > + if (atomic_dec_and_test(&sc->refcnt)) > + call_rcu(&sc->rcu_head, free_rx_sc_rcu); > +} > + > +static void free_rxsa(struct rcu_head *head) > +{ > + struct macsec_rx_sa *sa = container_of(head, struct macsec_rx_sa, rcu); > + > + crypto_free_aead(sa->key.tfm); > + free_percpu(sa->stats); > + macsec_rxsc_put(sa->sc); > + kfree(sa); > +} > + > +static void macsec_rxsa_put(struct macsec_rx_sa *sa) > +{ > + if (atomic_dec_and_test(&sa->refcnt)) > + call_rcu(&sa->rcu, free_rxsa); > +} > + > +static struct macsec_tx_sa *macsec_txsa_get(struct macsec_tx_sa __rcu *ptr) > +{ > + struct macsec_tx_sa *sa = rcu_dereference_bh(ptr); > + > + if (!sa || !sa->active) > + return NULL; > + > + if (!atomic_inc_not_zero(&sa->refcnt)) > + return NULL; > + > + return sa; > +} > + > +static void free_txsa(struct rcu_head *head) > +{ > + struct macsec_tx_sa *sa = container_of(head, struct macsec_tx_sa, rcu); > + > + crypto_free_aead(sa->key.tfm); > + free_percpu(sa->stats); > + kfree(sa); > +} > + > +static void macsec_txsa_put(struct macsec_tx_sa *sa) > +{ > + if (atomic_dec_and_test(&sa->refcnt)) > + call_rcu(&sa->rcu, free_txsa); > +} > + > +static struct macsec_cb *macsec_skb_cb(struct sk_buff *skb) > +{ > + BUILD_BUG_ON(sizeof(struct macsec_cb) > sizeof(skb->cb)); > + return (struct macsec_cb *)skb->cb; > +} > + > +#define MACSEC_PORT_ES (htons(0x0001)) > +#define MACSEC_PORT_SCB (0x0000) > +#define MACSEC_UNDEF_SCI ((__force sci_t)0xffffffffffffffffULL) > + > +#define DEFAULT_SAK_LEN 16 > +#define DEFAULT_SEND_SCI true > +#define DEFAULT_ENCRYPT false > +#define DEFAULT_ENCODING_SA 0 > + > +static sci_t make_sci(u8 *addr, __be16 port) > +{ > + sci_t sci; > + > + memcpy(&sci, addr, ETH_ALEN); > + memcpy(((char *)&sci) + ETH_ALEN, &port, sizeof(port)); > + > + return sci; > +} > + > +static sci_t macsec_frame_sci(struct macsec_eth_header *hdr, bool > sci_present) > +{ > + sci_t sci; > + > + if (sci_present) > + memcpy(&sci, hdr->secure_channel_id, > + sizeof(hdr->secure_channel_id)); > + else > + sci = make_sci(hdr->eth.h_source, MACSEC_PORT_ES); > + > + return sci; > +} > + > +static unsigned int macsec_sectag_len(bool sci_present) > +{ > + return MACSEC_TAG_LEN + (sci_present ? MACSEC_SCI_LEN : 0); > +} > + > +static unsigned int macsec_hdr_len(bool sci_present) > +{ > + return macsec_sectag_len(sci_present) + ETH_HLEN; > +} > + > +static unsigned int macsec_extra_len(bool sci_present) > +{ > + return macsec_sectag_len(sci_present) + sizeof(__be16); > +} > + > +/* Fill SecTAG according to IEEE 802.1AE-2006 10.5.3 */ > +static void macsec_fill_sectag(struct macsec_eth_header *h, > + const struct macsec_secy *secy, u32 pn) > +{ > + const struct macsec_tx_sc *tx_sc = &secy->tx_sc; > + > + memset(&h->tci_an, 0, macsec_sectag_len(tx_sc->send_sci)); > + h->eth.h_proto = htons(ETH_P_MACSEC); > + > + if (tx_sc->send_sci || > + (secy->n_rx_sc > 1 && !tx_sc->end_station && !tx_sc->scb)) { > + h->tci_an |= MACSEC_TCI_SC; > + memcpy(&h->secure_channel_id, &secy->sci, > + sizeof(h->secure_channel_id)); > + } else { > + if (tx_sc->end_station) > + h->tci_an |= MACSEC_TCI_ES; > + if (tx_sc->scb) > + h->tci_an |= MACSEC_TCI_SCB; > + } > + > + h->packet_number = htonl(pn); > + > + /* with GCM, C/E clear for !encrypt, both set for encrypt */ > + if (tx_sc->encrypt) > + h->tci_an |= MACSEC_TCI_CONFID; > + else if (secy->icv_len != DEFAULT_ICV_LEN) > + h->tci_an |= MACSEC_TCI_C; > + > + h->tci_an |= tx_sc->encoding_sa; > +} > + > +static void macsec_set_shortlen(struct macsec_eth_header *h, size_t data_len) > +{ > + if (data_len < MACSEC_SHORTLEN_THR) > + h->short_length = data_len; > +} > + > +/* minimum secure data length deemed "not short", see IEEE 802.1AE-2006 9.7 > */ > +#define MIN_NON_SHORT_LEN 48 > + > +/* validate MACsec packet according to IEEE 802.1AE-2006 9.12 */ > +static bool macsec_validate_skb(struct sk_buff *skb, u16 icv_len) > +{ > + struct macsec_eth_header *h = (struct macsec_eth_header *)skb->data; > + int len = skb->len - 2 * ETH_ALEN; > + int extra_len = macsec_extra_len(!!(h->tci_an & MACSEC_TCI_SC)) + > icv_len; > + > + /* a) It comprises at least 17 octets */ > + if (skb->len <= 16) > + return false; > + > + /* b) MACsec EtherType: already checked */ > + > + /* c) V bit is clear */ > + if (h->tci_an & MACSEC_TCI_VERSION) > + return false; > + > + /* d) ES or SCB => !SC */ > + if ((h->tci_an & MACSEC_TCI_ES || h->tci_an & MACSEC_TCI_SCB) && > + (h->tci_an & MACSEC_TCI_SC)) > + return false; > + > + /* e) Bits 7 and 8 of octet 4 of the SecTAG are clear */ > + if (h->unused) > + return false; > + > + /* rx.pn != 0 (figure 10-5) */ > + if (!h->packet_number) > + return false; > + > + /* length check, f) g) h) i) */ > + if (h->short_length) > + return len == extra_len + h->short_length; > + return len >= extra_len + MIN_NON_SHORT_LEN; > +} > + > +#define MACSEC_NEEDED_HEADROOM (macsec_extra_len(true)) > +#define MACSEC_NEEDED_TAILROOM MACSEC_MAX_ICV_LEN > + > +static void macsec_fill_iv(unsigned char *iv, sci_t sci, u32 pn) > +{ > + struct gcm_iv *gcm_iv = (struct gcm_iv *)iv; > + > + gcm_iv->sci = sci; > + gcm_iv->pn = htonl(pn); > +} > + > +static struct macsec_eth_header *macsec_ethhdr(struct sk_buff *skb) > +{ > + return (struct macsec_eth_header *)skb_mac_header(skb); > +} > + > +static u32 tx_sa_update_pn(struct macsec_tx_sa *tx_sa, struct macsec_secy > *secy) > +{ > + u32 pn; > + > + spin_lock_bh(&tx_sa->lock); > + pn = tx_sa->next_pn; > + > + tx_sa->next_pn++; > + if (tx_sa->next_pn == 0) { > + pr_debug("PN wrapped, transitionning to !oper\n"); > + tx_sa->active = false; > + if (secy->protect_frames) > + secy->operational = false; > + } > + spin_unlock_bh(&tx_sa->lock); > + > + return pn; > +} > + > +static void macsec_encrypt_finish(struct sk_buff *skb, struct net_device > *dev) > +{ > + struct macsec_dev *macsec = netdev_priv(dev); > + > + skb->dev = macsec->real_dev; > + skb_reset_mac_header(skb); > + skb->protocol = eth_hdr(skb)->h_proto; > +} > + > +static void macsec_count_tx(struct sk_buff *skb, struct macsec_tx_sc *tx_sc, > + struct macsec_tx_sa *tx_sa) > +{ > + struct pcpu_tx_sc_stats *txsc_stats = this_cpu_ptr(tx_sc->stats); > + > + u64_stats_update_begin(&txsc_stats->syncp); > + if (tx_sc->encrypt) { > + txsc_stats->stats.OutOctetsEncrypted += skb->len; > + txsc_stats->stats.OutPktsEncrypted++; > + this_cpu_inc(tx_sa->stats->OutPktsEncrypted); > + } else { > + txsc_stats->stats.OutOctetsProtected += skb->len; > + txsc_stats->stats.OutPktsProtected++; > + this_cpu_inc(tx_sa->stats->OutPktsProtected); > + } > + u64_stats_update_end(&txsc_stats->syncp); > +} > + > +static void count_tx(struct net_device *dev, int ret, int len) > +{ > + if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) { > + struct pcpu_sw_netstats *stats = this_cpu_ptr(dev->tstats); > + > + u64_stats_update_begin(&stats->syncp); > + stats->tx_packets++; > + stats->tx_bytes += len; > + u64_stats_update_end(&stats->syncp); > + } else { > + dev->stats.tx_dropped++; > + } > +} > + > +static void macsec_encrypt_done(struct crypto_async_request *base, int err) > +{ > + struct sk_buff *skb = base->data; > + struct net_device *dev = skb->dev; > + struct macsec_dev *macsec = macsec_priv(dev); > + struct macsec_tx_sa *sa = macsec_skb_cb(skb)->tx_sa; > + int len, ret; > + > + aead_request_free(macsec_skb_cb(skb)->req); > + > + rcu_read_lock_bh(); > + macsec_encrypt_finish(skb, dev); > + macsec_count_tx(skb, &macsec->secy.tx_sc, macsec_skb_cb(skb)->tx_sa); > + len = skb->len; > + ret = dev_queue_xmit(skb); > + count_tx(dev, ret, len); > + rcu_read_unlock_bh(); > + > + macsec_txsa_put(sa); > + dev_put(dev); > +} > + > +static struct sk_buff *macsec_encrypt(struct sk_buff *skb, > + struct net_device *dev) > +{ > + int ret; > + struct scatterlist sg[MAX_SKB_FRAGS + 1]; > + unsigned char iv[GCM_AES_IV_LEN]; > + struct ethhdr *eth; > + struct macsec_eth_header *hh; > + size_t unprotected_len; > + struct aead_request *req; > + struct macsec_secy *secy; > + struct macsec_tx_sc *tx_sc; > + struct macsec_tx_sa *tx_sa; > + struct macsec_dev *macsec = macsec_priv(dev); > + u32 pn; > + > + secy = &macsec->secy; > + tx_sc = &secy->tx_sc; > + > + /* 10.5.1 TX SA assignment */ > + tx_sa = macsec_txsa_get(tx_sc->sa[tx_sc->encoding_sa]); > + if (!tx_sa) { > + secy->operational = false; > + kfree_skb(skb); > + return ERR_PTR(-EINVAL); > + } > + > + if (unlikely(skb_headroom(skb) < MACSEC_NEEDED_HEADROOM || > + skb_tailroom(skb) < MACSEC_NEEDED_TAILROOM)) { > + struct sk_buff *nskb = skb_copy_expand(skb, > + MACSEC_NEEDED_HEADROOM, > + MACSEC_NEEDED_TAILROOM, > + GFP_ATOMIC); > + if (likely(nskb)) { > + consume_skb(skb); > + skb = nskb; > + } else { > + macsec_txsa_put(tx_sa); > + kfree_skb(skb); > + return ERR_PTR(-ENOMEM); > + } > + } else { > + skb = skb_unshare(skb, GFP_ATOMIC); > + if (!skb) { > + macsec_txsa_put(tx_sa); > + return ERR_PTR(-ENOMEM); > + } > + } > + > + unprotected_len = skb->len; > + eth = eth_hdr(skb); > + hh = (struct macsec_eth_header *)skb_push(skb, > macsec_extra_len(tx_sc->send_sci)); > + memmove(hh, eth, 2 * ETH_ALEN); > + > + pn = tx_sa_update_pn(tx_sa, secy); > + if (pn == 0) { > + macsec_txsa_put(tx_sa); > + kfree_skb(skb); > + return ERR_PTR(-ENOLINK); > + } > + macsec_fill_sectag(hh, secy, pn); > + macsec_set_shortlen(hh, unprotected_len - 2 * ETH_ALEN); > + > + macsec_fill_iv(iv, secy->sci, pn); > + > + skb_put(skb, secy->icv_len); > + > + if (skb->len - ETH_HLEN > macsec_priv(dev)->real_dev->mtu) { > + struct pcpu_secy_stats *secy_stats = > this_cpu_ptr(macsec->stats); > + > + u64_stats_update_begin(&secy_stats->syncp); > + secy_stats->stats.OutPktsTooLong++; > + u64_stats_update_end(&secy_stats->syncp); > + > + macsec_txsa_put(tx_sa); > + kfree_skb(skb); > + return ERR_PTR(-EINVAL); > + } > + > + req = aead_request_alloc(tx_sa->key.tfm, GFP_ATOMIC); > + if (!req) { > + macsec_txsa_put(tx_sa); > + kfree_skb(skb); > + return ERR_PTR(-ENOMEM); > + } > + > + sg_init_table(sg, MAX_SKB_FRAGS + 1); > + skb_to_sgvec(skb, sg, 0, skb->len); > + > + if (tx_sc->encrypt) { > + int len = skb->len - macsec_hdr_len(tx_sc->send_sci) - > + secy->icv_len; > + aead_request_set_crypt(req, sg, sg, len, iv); > + aead_request_set_ad(req, macsec_hdr_len(tx_sc->send_sci)); > + } else { > + aead_request_set_crypt(req, sg, sg, 0, iv); > + aead_request_set_ad(req, skb->len - secy->icv_len); > + } > + > + macsec_skb_cb(skb)->req = req; > + macsec_skb_cb(skb)->tx_sa = tx_sa; > + aead_request_set_callback(req, 0, macsec_encrypt_done, skb); > + > + dev_hold(skb->dev); > + ret = crypto_aead_encrypt(req); > + if (ret == -EINPROGRESS) { > + return ERR_PTR(ret); > + } else if (ret != 0) { > + dev_put(skb->dev); > + kfree_skb(skb); > + aead_request_free(req); > + macsec_txsa_put(tx_sa); > + return ERR_PTR(-EINVAL); > + } > + > + dev_put(skb->dev); > + aead_request_free(req); > + macsec_txsa_put(tx_sa); > + > + return skb; > +} > + > +static bool macsec_post_decrypt(struct sk_buff *skb, struct macsec_secy > *secy, u32 pn) > +{ > + struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; > + struct pcpu_rx_sc_stats *rxsc_stats = this_cpu_ptr(rx_sa->sc->stats); > + struct macsec_eth_header *hdr = macsec_ethhdr(skb); > + u32 lowest_pn = 0; > + > + spin_lock(&rx_sa->lock); > + if (rx_sa->next_pn >= secy->replay_window) > + lowest_pn = rx_sa->next_pn - secy->replay_window; > + > + /* Now perform replay protection check again > + * (see IEEE 802.1AE-2006 figure 10-5) > + */ > + if (secy->replay_protect && pn < lowest_pn) { > + spin_unlock(&rx_sa->lock); > + u64_stats_update_begin(&rxsc_stats->syncp); > + rxsc_stats->stats.InPktsLate++; > + u64_stats_update_end(&rxsc_stats->syncp); > + return false; > + } > + > + if (secy->validate_frames != MACSEC_VALIDATE_DISABLED) { > + u64_stats_update_begin(&rxsc_stats->syncp); > + if (hdr->tci_an & MACSEC_TCI_E) > + rxsc_stats->stats.InOctetsDecrypted += skb->len; > + else > + rxsc_stats->stats.InOctetsValidated += skb->len; > + u64_stats_update_end(&rxsc_stats->syncp); > + } > + > + if (!macsec_skb_cb(skb)->valid) { > + spin_unlock(&rx_sa->lock); > + > + /* 10.6.5 */ > + if (hdr->tci_an & MACSEC_TCI_C || > + secy->validate_frames == MACSEC_VALIDATE_STRICT) { > + u64_stats_update_begin(&rxsc_stats->syncp); > + rxsc_stats->stats.InPktsNotValid++; > + u64_stats_update_end(&rxsc_stats->syncp); > + return false; > + } > + > + u64_stats_update_begin(&rxsc_stats->syncp); > + if (secy->validate_frames == MACSEC_VALIDATE_CHECK) { > + rxsc_stats->stats.InPktsInvalid++; > + this_cpu_inc(rx_sa->stats->InPktsInvalid); > + } else if (pn < lowest_pn) { > + rxsc_stats->stats.InPktsDelayed++; > + } else { > + rxsc_stats->stats.InPktsUnchecked++; > + } > + u64_stats_update_end(&rxsc_stats->syncp); > + } else { > + u64_stats_update_begin(&rxsc_stats->syncp); > + if (pn < lowest_pn) { > + rxsc_stats->stats.InPktsDelayed++; > + } else { > + rxsc_stats->stats.InPktsOK++; > + this_cpu_inc(rx_sa->stats->InPktsOK); > + } > + u64_stats_update_end(&rxsc_stats->syncp); > + > + if (pn >= rx_sa->next_pn) > + rx_sa->next_pn = pn + 1; > + spin_unlock(&rx_sa->lock); > + } > + > + return true; > +} > + > +static void macsec_reset_skb(struct sk_buff *skb, struct net_device *dev) > +{ > + skb->pkt_type = PACKET_HOST; > + skb->protocol = eth_type_trans(skb, dev); > + > + skb_reset_network_header(skb); > + if (!skb_transport_header_was_set(skb)) > + skb_reset_transport_header(skb); > + skb_reset_mac_len(skb); > +} > + > +static void macsec_finalize_skb(struct sk_buff *skb, u8 icv_len, u8 hdr_len) > +{ > + memmove(skb->data + hdr_len, skb->data, 2 * ETH_ALEN); > + skb_pull(skb, hdr_len); > + pskb_trim_unique(skb, skb->len - icv_len); > +} > + > +static void count_rx(struct net_device *dev, int len) > +{ > + struct pcpu_sw_netstats *stats = this_cpu_ptr(dev->tstats); > + > + u64_stats_update_begin(&stats->syncp); > + stats->rx_packets++; > + stats->rx_bytes += len; > + u64_stats_update_end(&stats->syncp); > +} > + > +static void macsec_decrypt_done(struct crypto_async_request *base, int err) > +{ > + struct sk_buff *skb = base->data; > + struct net_device *dev = skb->dev; > + struct macsec_dev *macsec = macsec_priv(dev); > + struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; > + int len, ret; > + u32 pn; > + > + aead_request_free(macsec_skb_cb(skb)->req); > + > + rcu_read_lock_bh(); > + pn = ntohl(macsec_ethhdr(skb)->packet_number); > + if (!macsec_post_decrypt(skb, &macsec->secy, pn)) { > + rcu_read_unlock_bh(); > + kfree_skb(skb); > + goto out; > + } > + > + macsec_finalize_skb(skb, macsec->secy.icv_len, > + macsec_extra_len(macsec_skb_cb(skb)->has_sci)); > + macsec_reset_skb(skb, macsec->secy.netdev); > + > + len = skb->len; > + ret = netif_rx(skb); > + if (ret == NET_RX_SUCCESS) > + count_rx(dev, len); > + else > + macsec->secy.netdev->stats.rx_dropped++; > + > + rcu_read_unlock_bh(); > + > +out: > + macsec_rxsa_put(rx_sa); > + dev_put(dev); > + return; > +} > + > +static struct sk_buff *macsec_decrypt(struct sk_buff *skb, > + struct net_device *dev, > + struct macsec_rx_sa *rx_sa, > + sci_t sci, > + struct macsec_secy *secy) > +{ > + int ret; > + struct scatterlist sg[MAX_SKB_FRAGS + 1]; > + unsigned char iv[GCM_AES_IV_LEN]; > + struct aead_request *req; > + struct macsec_eth_header *hdr; > + u16 icv_len = secy->icv_len; > + > + macsec_skb_cb(skb)->valid = 0; Please use true/false for boolean variables. > + macsec_skb_cb(skb)->valid = 1; Likewise. > +static void handle_not_macsec(struct sk_buff *skb) > +{ > + struct macsec_rxh_data *rxd; > + struct macsec_dev *macsec; > + > + rcu_read_lock_bh(); "bh" should be implicit in this receive code path, so plain rcu_read_lock() should be sufficient. > +static rx_handler_result_t macsec_handle_frame(struct sk_buff **pskb) > +{ ... > + rcu_read_lock_bh(); Likewise. Otherwise looks really good to me.