commit:     ca0bbc3a2f4aa20a4945c343d948c7e739897b0a
Author:     Alice Ferrazzi <alicef <AT> gentoo <DOT> org>
AuthorDate: Sat Feb  4 11:36:38 2017 +0000
Commit:     Alice Ferrazzi <alicef <AT> gentoo <DOT> org>
CommitDate: Sat Feb  4 11:36:38 2017 +0000
URL:        https://gitweb.gentoo.org/proj/linux-patches.git/commit/?id=ca0bbc3a

Linux patch 4.9.8

 0000_README            |    4 +
 1007_linux-4.9.8.patch | 2048 ++++++++++++++++++++++++++++++++++++++++++++++++
 2 files changed, 2052 insertions(+)

diff --git a/0000_README b/0000_README
index edead0d..c9cccee 100644
--- a/0000_README
+++ b/0000_README
@@ -71,6 +71,10 @@ Patch:  1006_linux-4.9.7.patch
 From:   http://www.kernel.org
 Desc:   Linux 4.9.7
 
+Patch:  1007_linux-4.9.8.patch
+From:   http://www.kernel.org
+Desc:   Linux 4.9.8
+
 Patch:  1500_XATTR_USER_PREFIX.patch
 From:   https://bugs.gentoo.org/show_bug.cgi?id=470644
 Desc:   Support for namespace user.pax.* on tmpfs.

diff --git a/1007_linux-4.9.8.patch b/1007_linux-4.9.8.patch
new file mode 100644
index 0000000..7281821
--- /dev/null
+++ b/1007_linux-4.9.8.patch
@@ -0,0 +1,2048 @@
+diff --git a/Makefile b/Makefile
+index da704d903321..1130803ab93c 100644
+--- a/Makefile
++++ b/Makefile
+@@ -1,6 +1,6 @@
+ VERSION = 4
+ PATCHLEVEL = 9
+-SUBLEVEL = 7
++SUBLEVEL = 8
+ EXTRAVERSION =
+ NAME = Roaring Lionus
+ 
+diff --git a/drivers/net/ethernet/broadcom/bcmsysport.c 
b/drivers/net/ethernet/broadcom/bcmsysport.c
+index 25d1eb4933d0..be7ec5a76a54 100644
+--- a/drivers/net/ethernet/broadcom/bcmsysport.c
++++ b/drivers/net/ethernet/broadcom/bcmsysport.c
+@@ -710,11 +710,8 @@ static unsigned int __bcm_sysport_tx_reclaim(struct 
bcm_sysport_priv *priv,
+       unsigned int c_index, last_c_index, last_tx_cn, num_tx_cbs;
+       unsigned int pkts_compl = 0, bytes_compl = 0;
+       struct bcm_sysport_cb *cb;
+-      struct netdev_queue *txq;
+       u32 hw_ind;
+ 
+-      txq = netdev_get_tx_queue(ndev, ring->index);
+-
+       /* Compute how many descriptors have been processed since last call */
+       hw_ind = tdma_readl(priv, TDMA_DESC_RING_PROD_CONS_INDEX(ring->index));
+       c_index = (hw_ind >> RING_CONS_INDEX_SHIFT) & RING_CONS_INDEX_MASK;
+@@ -745,9 +742,6 @@ static unsigned int __bcm_sysport_tx_reclaim(struct 
bcm_sysport_priv *priv,
+ 
+       ring->c_index = c_index;
+ 
+-      if (netif_tx_queue_stopped(txq) && pkts_compl)
+-              netif_tx_wake_queue(txq);
+-
+       netif_dbg(priv, tx_done, ndev,
+                 "ring=%d c_index=%d pkts_compl=%d, bytes_compl=%d\n",
+                 ring->index, ring->c_index, pkts_compl, bytes_compl);
+@@ -759,16 +753,33 @@ static unsigned int __bcm_sysport_tx_reclaim(struct 
bcm_sysport_priv *priv,
+ static unsigned int bcm_sysport_tx_reclaim(struct bcm_sysport_priv *priv,
+                                          struct bcm_sysport_tx_ring *ring)
+ {
++      struct netdev_queue *txq;
+       unsigned int released;
+       unsigned long flags;
+ 
++      txq = netdev_get_tx_queue(priv->netdev, ring->index);
++
+       spin_lock_irqsave(&ring->lock, flags);
+       released = __bcm_sysport_tx_reclaim(priv, ring);
++      if (released)
++              netif_tx_wake_queue(txq);
++
+       spin_unlock_irqrestore(&ring->lock, flags);
+ 
+       return released;
+ }
+ 
++/* Locked version of the per-ring TX reclaim, but does not wake the queue */
++static void bcm_sysport_tx_clean(struct bcm_sysport_priv *priv,
++                               struct bcm_sysport_tx_ring *ring)
++{
++      unsigned long flags;
++
++      spin_lock_irqsave(&ring->lock, flags);
++      __bcm_sysport_tx_reclaim(priv, ring);
++      spin_unlock_irqrestore(&ring->lock, flags);
++}
++
+ static int bcm_sysport_tx_poll(struct napi_struct *napi, int budget)
+ {
+       struct bcm_sysport_tx_ring *ring =
+@@ -1253,7 +1264,7 @@ static void bcm_sysport_fini_tx_ring(struct 
bcm_sysport_priv *priv,
+       napi_disable(&ring->napi);
+       netif_napi_del(&ring->napi);
+ 
+-      bcm_sysport_tx_reclaim(priv, ring);
++      bcm_sysport_tx_clean(priv, ring);
+ 
+       kfree(ring->cbs);
+       ring->cbs = NULL;
+diff --git a/drivers/net/ethernet/mellanox/mlx4/en_netdev.c 
b/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
+index fb8bb027b69c..d223e7cb68ba 100644
+--- a/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
++++ b/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
+@@ -1740,8 +1740,11 @@ int mlx4_en_start_port(struct net_device *dev)
+       /* Process all completions if exist to prevent
+        * the queues freezing if they are full
+        */
+-      for (i = 0; i < priv->rx_ring_num; i++)
++      for (i = 0; i < priv->rx_ring_num; i++) {
++              local_bh_disable();
+               napi_schedule(&priv->rx_cq[i]->napi);
++              local_bh_enable();
++      }
+ 
+       netif_tx_start_all_queues(dev);
+       netif_device_attach(dev);
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c 
b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
+index 33495d88aeb2..e7b2158bb48a 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
+@@ -193,6 +193,9 @@ static inline bool mlx5e_rx_cache_put(struct mlx5e_rq *rq,
+               return false;
+       }
+ 
++      if (unlikely(page_is_pfmemalloc(dma_info->page)))
++              return false;
++
+       cache->page_cache[cache->tail] = *dma_info;
+       cache->tail = tail_next;
+       return true;
+diff --git a/drivers/net/ethernet/mellanox/mlxsw/pci.h 
b/drivers/net/ethernet/mellanox/mlxsw/pci.h
+index d942a3e6fa41..846fd4df7dab 100644
+--- a/drivers/net/ethernet/mellanox/mlxsw/pci.h
++++ b/drivers/net/ethernet/mellanox/mlxsw/pci.h
+@@ -211,21 +211,21 @@ MLXSW_ITEM32(pci, eqe, owner, 0x0C, 0, 1);
+ /* pci_eqe_cmd_token
+  * Command completion event - token
+  */
+-MLXSW_ITEM32(pci, eqe, cmd_token, 0x08, 16, 16);
++MLXSW_ITEM32(pci, eqe, cmd_token, 0x00, 16, 16);
+ 
+ /* pci_eqe_cmd_status
+  * Command completion event - status
+  */
+-MLXSW_ITEM32(pci, eqe, cmd_status, 0x08, 0, 8);
++MLXSW_ITEM32(pci, eqe, cmd_status, 0x00, 0, 8);
+ 
+ /* pci_eqe_cmd_out_param_h
+  * Command completion event - output parameter - higher part
+  */
+-MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x0C, 0, 32);
++MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x04, 0, 32);
+ 
+ /* pci_eqe_cmd_out_param_l
+  * Command completion event - output parameter - lower part
+  */
+-MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x10, 0, 32);
++MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x08, 0, 32);
+ 
+ #endif
+diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum.c 
b/drivers/net/ethernet/mellanox/mlxsw/spectrum.c
+index dda5761e91bc..f902c4d3de99 100644
+--- a/drivers/net/ethernet/mellanox/mlxsw/spectrum.c
++++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum.c
+@@ -684,6 +684,7 @@ static netdev_tx_t mlxsw_sp_port_xmit(struct sk_buff *skb,
+                       dev_kfree_skb_any(skb_orig);
+                       return NETDEV_TX_OK;
+               }
++              dev_consume_skb_any(skb_orig);
+       }
+ 
+       if (eth_skb_pad(skb)) {
+diff --git a/drivers/net/ethernet/mellanox/mlxsw/switchx2.c 
b/drivers/net/ethernet/mellanox/mlxsw/switchx2.c
+index 92bda8703f87..d548f0a55174 100644
+--- a/drivers/net/ethernet/mellanox/mlxsw/switchx2.c
++++ b/drivers/net/ethernet/mellanox/mlxsw/switchx2.c
+@@ -314,6 +314,7 @@ static netdev_tx_t mlxsw_sx_port_xmit(struct sk_buff *skb,
+                       dev_kfree_skb_any(skb_orig);
+                       return NETDEV_TX_OK;
+               }
++              dev_consume_skb_any(skb_orig);
+       }
+       mlxsw_sx_txhdr_construct(skb, &tx_info);
+       /* TX header is consumed by HW on the way so we shouldn't count its
+diff --git a/drivers/net/ethernet/renesas/ravb_main.c 
b/drivers/net/ethernet/renesas/ravb_main.c
+index d6a217874a8b..862f18ed6022 100644
+--- a/drivers/net/ethernet/renesas/ravb_main.c
++++ b/drivers/net/ethernet/renesas/ravb_main.c
+@@ -1508,6 +1508,19 @@ static netdev_tx_t ravb_start_xmit(struct sk_buff *skb, 
struct net_device *ndev)
+       buffer = PTR_ALIGN(priv->tx_align[q], DPTR_ALIGN) +
+                entry / NUM_TX_DESC * DPTR_ALIGN;
+       len = PTR_ALIGN(skb->data, DPTR_ALIGN) - skb->data;
++      /* Zero length DMA descriptors are problematic as they seem to
++       * terminate DMA transfers. Avoid them by simply using a length of
++       * DPTR_ALIGN (4) when skb data is aligned to DPTR_ALIGN.
++       *
++       * As skb is guaranteed to have at least ETH_ZLEN (60) bytes of
++       * data by the call to skb_put_padto() above this is safe with
++       * respect to both the length of the first DMA descriptor (len)
++       * overflowing the available data and the length of the second DMA
++       * descriptor (skb->len - len) being negative.
++       */
++      if (len == 0)
++              len = DPTR_ALIGN;
++
+       memcpy(buffer, skb->data, len);
+       dma_addr = dma_map_single(ndev->dev.parent, buffer, len, DMA_TO_DEVICE);
+       if (dma_mapping_error(ndev->dev.parent, dma_addr))
+diff --git a/drivers/net/hyperv/netvsc_drv.c b/drivers/net/hyperv/netvsc_drv.c
+index c9140c3aeb67..ff038e507fd6 100644
+--- a/drivers/net/hyperv/netvsc_drv.c
++++ b/drivers/net/hyperv/netvsc_drv.c
+@@ -659,6 +659,7 @@ int netvsc_recv_callback(struct hv_device *device_obj,
+        * policy filters on the host). Deliver these via the VF
+        * interface in the guest.
+        */
++      rcu_read_lock();
+       vf_netdev = rcu_dereference(net_device_ctx->vf_netdev);
+       if (vf_netdev && (vf_netdev->flags & IFF_UP))
+               net = vf_netdev;
+@@ -667,6 +668,7 @@ int netvsc_recv_callback(struct hv_device *device_obj,
+       skb = netvsc_alloc_recv_skb(net, packet, csum_info, *data, vlan_tci);
+       if (unlikely(!skb)) {
+               ++net->stats.rx_dropped;
++              rcu_read_unlock();
+               return NVSP_STAT_FAIL;
+       }
+ 
+@@ -696,6 +698,7 @@ int netvsc_recv_callback(struct hv_device *device_obj,
+        * TODO - use NAPI?
+        */
+       netif_rx(skb);
++      rcu_read_unlock();
+ 
+       return 0;
+ }
+diff --git a/drivers/net/macvtap.c b/drivers/net/macvtap.c
+index 7869b0651576..6f38daf2d978 100644
+--- a/drivers/net/macvtap.c
++++ b/drivers/net/macvtap.c
+@@ -827,7 +827,7 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q,
+                       return -EINVAL;
+ 
+               ret = virtio_net_hdr_from_skb(skb, &vnet_hdr,
+-                                            macvtap_is_little_endian(q));
++                                            macvtap_is_little_endian(q), 
true);
+               if (ret)
+                       BUG();
+ 
+diff --git a/drivers/net/phy/bcm63xx.c b/drivers/net/phy/bcm63xx.c
+index e741bf614c4e..b0492ef2cdaa 100644
+--- a/drivers/net/phy/bcm63xx.c
++++ b/drivers/net/phy/bcm63xx.c
+@@ -21,6 +21,23 @@ MODULE_DESCRIPTION("Broadcom 63xx internal PHY driver");
+ MODULE_AUTHOR("Maxime Bizon <[email protected]>");
+ MODULE_LICENSE("GPL");
+ 
++static int bcm63xx_config_intr(struct phy_device *phydev)
++{
++      int reg, err;
++
++      reg = phy_read(phydev, MII_BCM63XX_IR);
++      if (reg < 0)
++              return reg;
++
++      if (phydev->interrupts == PHY_INTERRUPT_ENABLED)
++              reg &= ~MII_BCM63XX_IR_GMASK;
++      else
++              reg |= MII_BCM63XX_IR_GMASK;
++
++      err = phy_write(phydev, MII_BCM63XX_IR, reg);
++      return err;
++}
++
+ static int bcm63xx_config_init(struct phy_device *phydev)
+ {
+       int reg, err;
+@@ -55,7 +72,7 @@ static struct phy_driver bcm63xx_driver[] = {
+       .config_aneg    = genphy_config_aneg,
+       .read_status    = genphy_read_status,
+       .ack_interrupt  = bcm_phy_ack_intr,
+-      .config_intr    = bcm_phy_config_intr,
++      .config_intr    = bcm63xx_config_intr,
+ }, {
+       /* same phy as above, with just a different OUI */
+       .phy_id         = 0x002bdc00,
+@@ -67,7 +84,7 @@ static struct phy_driver bcm63xx_driver[] = {
+       .config_aneg    = genphy_config_aneg,
+       .read_status    = genphy_read_status,
+       .ack_interrupt  = bcm_phy_ack_intr,
+-      .config_intr    = bcm_phy_config_intr,
++      .config_intr    = bcm63xx_config_intr,
+ } };
+ 
+ module_phy_driver(bcm63xx_driver);
+diff --git a/drivers/net/tun.c b/drivers/net/tun.c
+index db6acecabeaa..18402d79539e 100644
+--- a/drivers/net/tun.c
++++ b/drivers/net/tun.c
+@@ -1374,7 +1374,7 @@ static ssize_t tun_put_user(struct tun_struct *tun,
+                       return -EINVAL;
+ 
+               ret = virtio_net_hdr_from_skb(skb, &gso,
+-                                            tun_is_little_endian(tun));
++                                            tun_is_little_endian(tun), true);
+               if (ret) {
+                       struct skb_shared_info *sinfo = skb_shinfo(skb);
+                       pr_err("unexpected GSO type: "
+diff --git a/drivers/net/usb/cdc_ether.c b/drivers/net/usb/cdc_ether.c
+index dd623f674487..b82be816256c 100644
+--- a/drivers/net/usb/cdc_ether.c
++++ b/drivers/net/usb/cdc_ether.c
+@@ -531,6 +531,7 @@ static const struct driver_info wwan_info = {
+ #define SAMSUNG_VENDOR_ID     0x04e8
+ #define LENOVO_VENDOR_ID      0x17ef
+ #define NVIDIA_VENDOR_ID      0x0955
++#define HP_VENDOR_ID          0x03f0
+ 
+ static const struct usb_device_id     products[] = {
+ /* BLACKLIST !!
+@@ -677,6 +678,13 @@ static const struct usb_device_id products[] = {
+       .driver_info = 0,
+ },
+ 
++/* HP lt2523 (Novatel E371) - handled by qmi_wwan */
++{
++      USB_DEVICE_AND_INTERFACE_INFO(HP_VENDOR_ID, 0x421d, USB_CLASS_COMM,
++                                    USB_CDC_SUBCLASS_ETHERNET, 
USB_CDC_PROTO_NONE),
++      .driver_info = 0,
++},
++
+ /* AnyDATA ADU960S - handled by qmi_wwan */
+ {
+       USB_DEVICE_AND_INTERFACE_INFO(0x16d5, 0x650a, USB_CLASS_COMM,
+diff --git a/drivers/net/usb/qmi_wwan.c b/drivers/net/usb/qmi_wwan.c
+index 6fe1cdb0174f..24d5272cdce5 100644
+--- a/drivers/net/usb/qmi_wwan.c
++++ b/drivers/net/usb/qmi_wwan.c
+@@ -654,6 +654,13 @@ static const struct usb_device_id products[] = {
+                                             USB_CDC_PROTO_NONE),
+               .driver_info        = (unsigned long)&qmi_wwan_info,
+       },
++      {       /* HP lt2523 (Novatel E371) */
++              USB_DEVICE_AND_INTERFACE_INFO(0x03f0, 0x421d,
++                                            USB_CLASS_COMM,
++                                            USB_CDC_SUBCLASS_ETHERNET,
++                                            USB_CDC_PROTO_NONE),
++              .driver_info        = (unsigned long)&qmi_wwan_info,
++      },
+       {       /* HP lt4112 LTE/HSPA+ Gobi 4G Module (Huawei me906e) */
+               USB_DEVICE_AND_INTERFACE_INFO(0x03f0, 0x581d, 
USB_CLASS_VENDOR_SPEC, 1, 7),
+               .driver_info = (unsigned long)&qmi_wwan_info,
+diff --git a/drivers/net/usb/r8152.c b/drivers/net/usb/r8152.c
+index 4b5cb162442b..90b426c5ffce 100644
+--- a/drivers/net/usb/r8152.c
++++ b/drivers/net/usb/r8152.c
+@@ -32,7 +32,7 @@
+ #define NETNEXT_VERSION               "08"
+ 
+ /* Information for net */
+-#define NET_VERSION           "6"
++#define NET_VERSION           "7"
+ 
+ #define DRIVER_VERSION                "v1." NETNEXT_VERSION "." NET_VERSION
+ #define DRIVER_AUTHOR "Realtek linux nic maintainers <[email protected]>"
+@@ -1730,7 +1730,7 @@ static u8 r8152_rx_csum(struct r8152 *tp, struct rx_desc 
*rx_desc)
+       u8 checksum = CHECKSUM_NONE;
+       u32 opts2, opts3;
+ 
+-      if (tp->version == RTL_VER_01 || tp->version == RTL_VER_02)
++      if (!(tp->netdev->features & NETIF_F_RXCSUM))
+               goto return_result;
+ 
+       opts2 = le32_to_cpu(rx_desc->opts2);
+@@ -3572,6 +3572,8 @@ static bool delay_autosuspend(struct r8152 *tp)
+        */
+       if (!sw_linking && tp->rtl_ops.in_nway(tp))
+               return true;
++      else if (!skb_queue_empty(&tp->tx_queue))
++              return true;
+       else
+               return false;
+ }
+@@ -4358,6 +4360,11 @@ static int rtl8152_probe(struct usb_interface *intf,
+                               NETIF_F_HIGHDMA | NETIF_F_FRAGLIST |
+                               NETIF_F_IPV6_CSUM | NETIF_F_TSO6;
+ 
++      if (tp->version == RTL_VER_01) {
++              netdev->features &= ~NETIF_F_RXCSUM;
++              netdev->hw_features &= ~NETIF_F_RXCSUM;
++      }
++
+       netdev->ethtool_ops = &ops;
+       netif_set_gso_max_size(netdev, RTL_LIMITED_TSO_SIZE);
+ 
+diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
+index cbf1c613c67a..51fc0c33a62f 100644
+--- a/drivers/net/virtio_net.c
++++ b/drivers/net/virtio_net.c
+@@ -840,7 +840,7 @@ static int xmit_skb(struct send_queue *sq, struct sk_buff 
*skb)
+               hdr = skb_vnet_hdr(skb);
+ 
+       if (virtio_net_hdr_from_skb(skb, &hdr->hdr,
+-                                  virtio_is_little_endian(vi->vdev)))
++                                  virtio_is_little_endian(vi->vdev), false))
+               BUG();
+ 
+       if (vi->mergeable_rx_bufs)
+diff --git a/drivers/net/vxlan.c b/drivers/net/vxlan.c
+index 2ba01ca02c9c..0fafaa9d903b 100644
+--- a/drivers/net/vxlan.c
++++ b/drivers/net/vxlan.c
+@@ -2887,7 +2887,7 @@ static int vxlan_dev_configure(struct net *src_net, 
struct net_device *dev,
+       memcpy(&vxlan->cfg, conf, sizeof(*conf));
+       if (!vxlan->cfg.dst_port) {
+               if (conf->flags & VXLAN_F_GPE)
+-                      vxlan->cfg.dst_port = 4790; /* IANA assigned VXLAN-GPE 
port */
++                      vxlan->cfg.dst_port = htons(4790); /* IANA VXLAN-GPE 
port */
+               else
+                       vxlan->cfg.dst_port = default_port;
+       }
+diff --git a/fs/xfs/libxfs/xfs_alloc.c b/fs/xfs/libxfs/xfs_alloc.c
+index 5050056a0b06..9f06a211e157 100644
+--- a/fs/xfs/libxfs/xfs_alloc.c
++++ b/fs/xfs/libxfs/xfs_alloc.c
+@@ -95,10 +95,7 @@ unsigned int
+ xfs_alloc_set_aside(
+       struct xfs_mount        *mp)
+ {
+-      unsigned int            blocks;
+-
+-      blocks = 4 + (mp->m_sb.sb_agcount * XFS_ALLOC_AGFL_RESERVE);
+-      return blocks;
++      return mp->m_sb.sb_agcount * (XFS_ALLOC_AGFL_RESERVE + 4);
+ }
+ 
+ /*
+@@ -365,36 +362,12 @@ xfs_alloc_fix_len(
+               return;
+       ASSERT(rlen >= args->minlen && rlen <= args->maxlen);
+       ASSERT(rlen % args->prod == args->mod);
++      ASSERT(args->pag->pagf_freeblks + args->pag->pagf_flcount >=
++              rlen + args->minleft);
+       args->len = rlen;
+ }
+ 
+ /*
+- * Fix up length if there is too little space left in the a.g.
+- * Return 1 if ok, 0 if too little, should give up.
+- */
+-STATIC int
+-xfs_alloc_fix_minleft(
+-      xfs_alloc_arg_t *args)          /* allocation argument structure */
+-{
+-      xfs_agf_t       *agf;           /* a.g. freelist header */
+-      int             diff;           /* free space difference */
+-
+-      if (args->minleft == 0)
+-              return 1;
+-      agf = XFS_BUF_TO_AGF(args->agbp);
+-      diff = be32_to_cpu(agf->agf_freeblks)
+-              - args->len - args->minleft;
+-      if (diff >= 0)
+-              return 1;
+-      args->len += diff;              /* shrink the allocated space */
+-      /* casts to (int) catch length underflows */
+-      if ((int)args->len >= (int)args->minlen)
+-              return 1;
+-      args->agbno = NULLAGBLOCK;
+-      return 0;
+-}
+-
+-/*
+  * Update the two btrees, logically removing from freespace the extent
+  * starting at rbno, rlen blocks.  The extent is contained within the
+  * actual (current) free extent fbno for flen blocks.
+@@ -689,8 +662,6 @@ xfs_alloc_ag_vextent(
+       xfs_alloc_arg_t *args)  /* argument structure for allocation */
+ {
+       int             error=0;
+-      xfs_extlen_t    reservation;
+-      xfs_extlen_t    oldmax;
+ 
+       ASSERT(args->minlen > 0);
+       ASSERT(args->maxlen > 0);
+@@ -699,20 +670,6 @@ xfs_alloc_ag_vextent(
+       ASSERT(args->alignment > 0);
+ 
+       /*
+-       * Clamp maxlen to the amount of free space minus any reservations
+-       * that have been made.
+-       */
+-      oldmax = args->maxlen;
+-      reservation = xfs_ag_resv_needed(args->pag, args->resv);
+-      if (args->maxlen > args->pag->pagf_freeblks - reservation)
+-              args->maxlen = args->pag->pagf_freeblks - reservation;
+-      if (args->maxlen == 0) {
+-              args->agbno = NULLAGBLOCK;
+-              args->maxlen = oldmax;
+-              return 0;
+-      }
+-
+-      /*
+        * Branch to correct routine based on the type.
+        */
+       args->wasfromfl = 0;
+@@ -731,8 +688,6 @@ xfs_alloc_ag_vextent(
+               /* NOTREACHED */
+       }
+ 
+-      args->maxlen = oldmax;
+-
+       if (error || args->agbno == NULLAGBLOCK)
+               return error;
+ 
+@@ -841,9 +796,6 @@ xfs_alloc_ag_vextent_exact(
+       args->len = XFS_AGBLOCK_MIN(tend, args->agbno + args->maxlen)
+                                               - args->agbno;
+       xfs_alloc_fix_len(args);
+-      if (!xfs_alloc_fix_minleft(args))
+-              goto not_found;
+-
+       ASSERT(args->agbno + args->len <= tend);
+ 
+       /*
+@@ -1149,12 +1101,7 @@ xfs_alloc_ag_vextent_near(
+               XFS_WANT_CORRUPTED_GOTO(args->mp, i == 1, error0);
+               ASSERT(ltbno + ltlen <= 
be32_to_cpu(XFS_BUF_TO_AGF(args->agbp)->agf_length));
+               args->len = blen;
+-              if (!xfs_alloc_fix_minleft(args)) {
+-                      xfs_btree_del_cursor(cnt_cur, XFS_BTREE_NOERROR);
+-                      trace_xfs_alloc_near_nominleft(args);
+-                      return 0;
+-              }
+-              blen = args->len;
++
+               /*
+                * We are allocating starting at bnew for blen blocks.
+                */
+@@ -1346,12 +1293,6 @@ xfs_alloc_ag_vextent_near(
+        */
+       args->len = XFS_EXTLEN_MIN(ltlena, args->maxlen);
+       xfs_alloc_fix_len(args);
+-      if (!xfs_alloc_fix_minleft(args)) {
+-              trace_xfs_alloc_near_nominleft(args);
+-              xfs_btree_del_cursor(bno_cur_lt, XFS_BTREE_NOERROR);
+-              xfs_btree_del_cursor(cnt_cur, XFS_BTREE_NOERROR);
+-              return 0;
+-      }
+       rlen = args->len;
+       (void)xfs_alloc_compute_diff(args->agbno, rlen, args->alignment,
+                                    args->datatype, ltbnoa, ltlena, &ltnew);
+@@ -1553,8 +1494,6 @@ xfs_alloc_ag_vextent_size(
+       }
+       xfs_alloc_fix_len(args);
+ 
+-      if (!xfs_alloc_fix_minleft(args))
+-              goto out_nominleft;
+       rlen = args->len;
+       XFS_WANT_CORRUPTED_GOTO(args->mp, rlen <= flen, error0);
+       /*
+@@ -2056,7 +1995,7 @@ xfs_alloc_space_available(
+       int                     flags)
+ {
+       struct xfs_perag        *pag = args->pag;
+-      xfs_extlen_t            longest;
++      xfs_extlen_t            alloc_len, longest;
+       xfs_extlen_t            reservation; /* blocks that are still reserved 
*/
+       int                     available;
+ 
+@@ -2066,17 +2005,28 @@ xfs_alloc_space_available(
+       reservation = xfs_ag_resv_needed(pag, args->resv);
+ 
+       /* do we have enough contiguous free space for the allocation? */
++      alloc_len = args->minlen + (args->alignment - 1) + args->minalignslop;
+       longest = xfs_alloc_longest_free_extent(args->mp, pag, min_free,
+                       reservation);
+-      if ((args->minlen + args->alignment + args->minalignslop - 1) > longest)
++      if (longest < alloc_len)
+               return false;
+ 
+       /* do we have enough free space remaining for the allocation? */
+       available = (int)(pag->pagf_freeblks + pag->pagf_flcount -
+-                        reservation - min_free - args->total);
+-      if (available < (int)args->minleft || available <= 0)
++                        reservation - min_free - args->minleft);
++      if (available < (int)max(args->total, alloc_len))
+               return false;
+ 
++      /*
++       * Clamp maxlen to the amount of free space available for the actual
++       * extent allocation.
++       */
++      if (available < (int)args->maxlen && !(flags & XFS_ALLOC_FLAG_CHECK)) {
++              args->maxlen = available;
++              ASSERT(args->maxlen > 0);
++              ASSERT(args->maxlen >= args->minlen);
++      }
++
+       return true;
+ }
+ 
+@@ -2122,7 +2072,8 @@ xfs_alloc_fix_freelist(
+       }
+ 
+       need = xfs_alloc_min_freelist(mp, pag);
+-      if (!xfs_alloc_space_available(args, need, flags))
++      if (!xfs_alloc_space_available(args, need, flags |
++                      XFS_ALLOC_FLAG_CHECK))
+               goto out_agbp_relse;
+ 
+       /*
+@@ -2638,12 +2589,10 @@ xfs_alloc_vextent(
+       xfs_agblock_t   agsize; /* allocation group size */
+       int             error;
+       int             flags;  /* XFS_ALLOC_FLAG_... locking flags */
+-      xfs_extlen_t    minleft;/* minimum left value, temp copy */
+       xfs_mount_t     *mp;    /* mount structure pointer */
+       xfs_agnumber_t  sagno;  /* starting allocation group number */
+       xfs_alloctype_t type;   /* input allocation type */
+       int             bump_rotor = 0;
+-      int             no_min = 0;
+       xfs_agnumber_t  rotorstep = xfs_rotorstep; /* inode32 agf stepper */
+ 
+       mp = args->mp;
+@@ -2672,7 +2621,6 @@ xfs_alloc_vextent(
+               trace_xfs_alloc_vextent_badargs(args);
+               return 0;
+       }
+-      minleft = args->minleft;
+ 
+       switch (type) {
+       case XFS_ALLOCTYPE_THIS_AG:
+@@ -2683,9 +2631,7 @@ xfs_alloc_vextent(
+                */
+               args->agno = XFS_FSB_TO_AGNO(mp, args->fsbno);
+               args->pag = xfs_perag_get(mp, args->agno);
+-              args->minleft = 0;
+               error = xfs_alloc_fix_freelist(args, 0);
+-              args->minleft = minleft;
+               if (error) {
+                       trace_xfs_alloc_vextent_nofix(args);
+                       goto error0;
+@@ -2750,9 +2696,7 @@ xfs_alloc_vextent(
+                */
+               for (;;) {
+                       args->pag = xfs_perag_get(mp, args->agno);
+-                      if (no_min) args->minleft = 0;
+                       error = xfs_alloc_fix_freelist(args, flags);
+-                      args->minleft = minleft;
+                       if (error) {
+                               trace_xfs_alloc_vextent_nofix(args);
+                               goto error0;
+@@ -2792,20 +2736,17 @@ xfs_alloc_vextent(
+                        * or switch to non-trylock mode.
+                        */
+                       if (args->agno == sagno) {
+-                              if (no_min == 1) {
++                              if (flags == 0) {
+                                       args->agbno = NULLAGBLOCK;
+                                       trace_xfs_alloc_vextent_allfailed(args);
+                                       break;
+                               }
+-                              if (flags == 0) {
+-                                      no_min = 1;
+-                              } else {
+-                                      flags = 0;
+-                                      if (type == XFS_ALLOCTYPE_START_BNO) {
+-                                              args->agbno = 
XFS_FSB_TO_AGBNO(mp,
+-                                                      args->fsbno);
+-                                              args->type = 
XFS_ALLOCTYPE_NEAR_BNO;
+-                                      }
++
++                              flags = 0;
++                              if (type == XFS_ALLOCTYPE_START_BNO) {
++                                      args->agbno = XFS_FSB_TO_AGBNO(mp,
++                                              args->fsbno);
++                                      args->type = XFS_ALLOCTYPE_NEAR_BNO;
+                               }
+                       }
+                       xfs_perag_put(args->pag);
+diff --git a/fs/xfs/libxfs/xfs_alloc.h b/fs/xfs/libxfs/xfs_alloc.h
+index 7c404a6b0ae3..1d0f48a501a3 100644
+--- a/fs/xfs/libxfs/xfs_alloc.h
++++ b/fs/xfs/libxfs/xfs_alloc.h
+@@ -56,7 +56,7 @@ typedef unsigned int xfs_alloctype_t;
+ #define       XFS_ALLOC_FLAG_FREEING  0x00000002  /* indicate caller is 
freeing extents*/
+ #define       XFS_ALLOC_FLAG_NORMAP   0x00000004  /* don't modify the rmapbt 
*/
+ #define       XFS_ALLOC_FLAG_NOSHRINK 0x00000008  /* don't shrink the 
freelist */
+-
++#define       XFS_ALLOC_FLAG_CHECK    0x00000010  /* test only, don't modify 
args */
+ 
+ /*
+  * Argument structure for xfs_alloc routines.
+diff --git a/fs/xfs/libxfs/xfs_attr.c b/fs/xfs/libxfs/xfs_attr.c
+index af1ecb19121e..6622d46ddec3 100644
+--- a/fs/xfs/libxfs/xfs_attr.c
++++ b/fs/xfs/libxfs/xfs_attr.c
+@@ -131,9 +131,6 @@ xfs_attr_get(
+       if (XFS_FORCED_SHUTDOWN(ip->i_mount))
+               return -EIO;
+ 
+-      if (!xfs_inode_hasattr(ip))
+-              return -ENOATTR;
+-
+       error = xfs_attr_args_init(&args, ip, name, flags);
+       if (error)
+               return error;
+@@ -392,9 +389,6 @@ xfs_attr_remove(
+       if (XFS_FORCED_SHUTDOWN(dp->i_mount))
+               return -EIO;
+ 
+-      if (!xfs_inode_hasattr(dp))
+-              return -ENOATTR;
+-
+       error = xfs_attr_args_init(&args, dp, name, flags);
+       if (error)
+               return error;
+diff --git a/fs/xfs/libxfs/xfs_bmap.c b/fs/xfs/libxfs/xfs_bmap.c
+index 89d727b659fc..f52fd63fce19 100644
+--- a/fs/xfs/libxfs/xfs_bmap.c
++++ b/fs/xfs/libxfs/xfs_bmap.c
+@@ -3720,7 +3720,7 @@ xfs_bmap_btalloc(
+               align = xfs_get_cowextsz_hint(ap->ip);
+       else if (xfs_alloc_is_userdata(ap->datatype))
+               align = xfs_get_extsz_hint(ap->ip);
+-      if (unlikely(align)) {
++      if (align) {
+               error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev,
+                                               align, 0, ap->eof, 0, ap->conv,
+                                               &ap->offset, &ap->length);
+@@ -3792,7 +3792,7 @@ xfs_bmap_btalloc(
+               args.minlen = ap->minlen;
+       }
+       /* apply extent size hints if obtained earlier */
+-      if (unlikely(align)) {
++      if (align) {
+               args.prod = align;
+               if ((args.mod = (xfs_extlen_t)do_mod(ap->offset, args.prod)))
+                       args.mod = (xfs_extlen_t)(args.prod - args.mod);
+@@ -3903,7 +3903,6 @@ xfs_bmap_btalloc(
+               args.fsbno = 0;
+               args.type = XFS_ALLOCTYPE_FIRST_AG;
+               args.total = ap->minlen;
+-              args.minleft = 0;
+               if ((error = xfs_alloc_vextent(&args)))
+                       return error;
+               ap->dfops->dop_low = true;
+@@ -4437,8 +4436,6 @@ xfs_bmapi_allocate(
+       if (error)
+               return error;
+ 
+-      if (bma->dfops->dop_low)
+-              bma->minleft = 0;
+       if (bma->cur)
+               bma->cur->bc_private.b.firstblock = *bma->firstblock;
+       if (bma->blkno == NULLFSBLOCK)
+@@ -4610,8 +4607,6 @@ xfs_bmapi_write(
+       int                     n;              /* current extent index */
+       xfs_fileoff_t           obno;           /* old block number (offset) */
+       int                     whichfork;      /* data or attr fork */
+-      char                    inhole;         /* current location is hole in 
file */
+-      char                    wasdelay;       /* old extent was delayed */
+ 
+ #ifdef DEBUG
+       xfs_fileoff_t           orig_bno;       /* original block number value 
*/
+@@ -4697,22 +4692,44 @@ xfs_bmapi_write(
+       bma.firstblock = firstblock;
+ 
+       while (bno < end && n < *nmap) {
+-              inhole = eof || bma.got.br_startoff > bno;
+-              wasdelay = !inhole && isnullstartblock(bma.got.br_startblock);
++              bool                    need_alloc = false, wasdelay = false;
+ 
+-              /*
+-               * Make sure we only reflink into a hole.
+-               */
+-              if (flags & XFS_BMAPI_REMAP)
+-                      ASSERT(inhole);
+-              if (flags & XFS_BMAPI_COWFORK)
+-                      ASSERT(!inhole);
++              /* in hole or beyoned EOF? */
++              if (eof || bma.got.br_startoff > bno) {
++                      if (flags & XFS_BMAPI_DELALLOC) {
++                              /*
++                               * For the COW fork we can reasonably get a
++                               * request for converting an extent that races
++                               * with other threads already having converted
++                               * part of it, as there converting COW to
++                               * regular blocks is not protected using the
++                               * IOLOCK.
++                               */
++                              ASSERT(flags & XFS_BMAPI_COWFORK);
++                              if (!(flags & XFS_BMAPI_COWFORK)) {
++                                      error = -EIO;
++                                      goto error0;
++                              }
++
++                              if (eof || bno >= end)
++                                      break;
++                      } else {
++                              need_alloc = true;
++                      }
++              } else {
++                      /*
++                       * Make sure we only reflink into a hole.
++                       */
++                      ASSERT(!(flags & XFS_BMAPI_REMAP));
++                      if (isnullstartblock(bma.got.br_startblock))
++                              wasdelay = true;
++              }
+ 
+               /*
+                * First, deal with the hole before the allocated space
+                * that we found, if any.
+                */
+-              if (inhole || wasdelay) {
++              if (need_alloc || wasdelay) {
+                       bma.eof = eof;
+                       bma.conv = !!(flags & XFS_BMAPI_CONVERT);
+                       bma.wasdel = wasdelay;
+diff --git a/fs/xfs/libxfs/xfs_bmap.h b/fs/xfs/libxfs/xfs_bmap.h
+index d6d175a4fdec..e7d40b39f18f 100644
+--- a/fs/xfs/libxfs/xfs_bmap.h
++++ b/fs/xfs/libxfs/xfs_bmap.h
+@@ -110,6 +110,9 @@ struct xfs_extent_free_item
+ /* Map something in the CoW fork. */
+ #define XFS_BMAPI_COWFORK     0x200
+ 
++/* Only convert delalloc space, don't allocate entirely new extents */
++#define XFS_BMAPI_DELALLOC    0x400
++
+ #define XFS_BMAPI_FLAGS \
+       { XFS_BMAPI_ENTIRE,     "ENTIRE" }, \
+       { XFS_BMAPI_METADATA,   "METADATA" }, \
+@@ -120,7 +123,8 @@ struct xfs_extent_free_item
+       { XFS_BMAPI_CONVERT,    "CONVERT" }, \
+       { XFS_BMAPI_ZERO,       "ZERO" }, \
+       { XFS_BMAPI_REMAP,      "REMAP" }, \
+-      { XFS_BMAPI_COWFORK,    "COWFORK" }
++      { XFS_BMAPI_COWFORK,    "COWFORK" }, \
++      { XFS_BMAPI_DELALLOC,   "DELALLOC" }
+ 
+ 
+ static inline int xfs_bmapi_aflag(int w)
+diff --git a/fs/xfs/libxfs/xfs_bmap_btree.c b/fs/xfs/libxfs/xfs_bmap_btree.c
+index 049fa597ae91..f76c1693ff01 100644
+--- a/fs/xfs/libxfs/xfs_bmap_btree.c
++++ b/fs/xfs/libxfs/xfs_bmap_btree.c
+@@ -502,12 +502,11 @@ xfs_bmbt_alloc_block(
+       if (args.fsbno == NULLFSBLOCK && args.minleft) {
+               /*
+                * Could not find an AG with enough free space to satisfy
+-               * a full btree split.  Try again without minleft and if
++               * a full btree split.  Try again and if
+                * successful activate the lowspace algorithm.
+                */
+               args.fsbno = 0;
+               args.type = XFS_ALLOCTYPE_FIRST_AG;
+-              args.minleft = 0;
+               error = xfs_alloc_vextent(&args);
+               if (error)
+                       goto error0;
+diff --git a/fs/xfs/libxfs/xfs_dir2.c b/fs/xfs/libxfs/xfs_dir2.c
+index 20a96dd5af7e..7825d78d4587 100644
+--- a/fs/xfs/libxfs/xfs_dir2.c
++++ b/fs/xfs/libxfs/xfs_dir2.c
+@@ -36,21 +36,29 @@
+ struct xfs_name xfs_name_dotdot = { (unsigned char *)"..", 2, XFS_DIR3_FT_DIR 
};
+ 
+ /*
+- * @mode, if set, indicates that the type field needs to be set up.
+- * This uses the transformation from file mode to DT_* as defined in 
linux/fs.h
+- * for file type specification. This will be propagated into the directory
+- * structure if appropriate for the given operation and filesystem config.
++ * Convert inode mode to directory entry filetype
+  */
+-const unsigned char xfs_mode_to_ftype[S_IFMT >> S_SHIFT] = {
+-      [0]                     = XFS_DIR3_FT_UNKNOWN,
+-      [S_IFREG >> S_SHIFT]    = XFS_DIR3_FT_REG_FILE,
+-      [S_IFDIR >> S_SHIFT]    = XFS_DIR3_FT_DIR,
+-      [S_IFCHR >> S_SHIFT]    = XFS_DIR3_FT_CHRDEV,
+-      [S_IFBLK >> S_SHIFT]    = XFS_DIR3_FT_BLKDEV,
+-      [S_IFIFO >> S_SHIFT]    = XFS_DIR3_FT_FIFO,
+-      [S_IFSOCK >> S_SHIFT]   = XFS_DIR3_FT_SOCK,
+-      [S_IFLNK >> S_SHIFT]    = XFS_DIR3_FT_SYMLINK,
+-};
++unsigned char xfs_mode_to_ftype(int mode)
++{
++      switch (mode & S_IFMT) {
++      case S_IFREG:
++              return XFS_DIR3_FT_REG_FILE;
++      case S_IFDIR:
++              return XFS_DIR3_FT_DIR;
++      case S_IFCHR:
++              return XFS_DIR3_FT_CHRDEV;
++      case S_IFBLK:
++              return XFS_DIR3_FT_BLKDEV;
++      case S_IFIFO:
++              return XFS_DIR3_FT_FIFO;
++      case S_IFSOCK:
++              return XFS_DIR3_FT_SOCK;
++      case S_IFLNK:
++              return XFS_DIR3_FT_SYMLINK;
++      default:
++              return XFS_DIR3_FT_UNKNOWN;
++      }
++}
+ 
+ /*
+  * ASCII case-insensitive (ie. A-Z) support for directories that was
+@@ -631,7 +639,8 @@ xfs_dir2_isblock(
+       if ((rval = xfs_bmap_last_offset(args->dp, &last, XFS_DATA_FORK)))
+               return rval;
+       rval = XFS_FSB_TO_B(args->dp->i_mount, last) == args->geo->blksize;
+-      ASSERT(rval == 0 || args->dp->i_d.di_size == args->geo->blksize);
++      if (rval != 0 && args->dp->i_d.di_size != args->geo->blksize)
++              return -EFSCORRUPTED;
+       *vp = rval;
+       return 0;
+ }
+diff --git a/fs/xfs/libxfs/xfs_dir2.h b/fs/xfs/libxfs/xfs_dir2.h
+index becc926c3e3d..ae0d55bf6500 100644
+--- a/fs/xfs/libxfs/xfs_dir2.h
++++ b/fs/xfs/libxfs/xfs_dir2.h
+@@ -18,6 +18,9 @@
+ #ifndef __XFS_DIR2_H__
+ #define __XFS_DIR2_H__
+ 
++#include "xfs_da_format.h"
++#include "xfs_da_btree.h"
++
+ struct xfs_defer_ops;
+ struct xfs_da_args;
+ struct xfs_inode;
+@@ -32,10 +35,9 @@ struct xfs_dir2_data_unused;
+ extern struct xfs_name        xfs_name_dotdot;
+ 
+ /*
+- * directory filetype conversion tables.
++ * Convert inode mode to directory entry filetype
+  */
+-#define S_SHIFT 12
+-extern const unsigned char xfs_mode_to_ftype[];
++extern unsigned char xfs_mode_to_ftype(int mode);
+ 
+ /*
+  * directory operations vector for encode/decode routines
+diff --git a/fs/xfs/libxfs/xfs_inode_buf.c b/fs/xfs/libxfs/xfs_inode_buf.c
+index c906e50515f0..37ee7f01a35d 100644
+--- a/fs/xfs/libxfs/xfs_inode_buf.c
++++ b/fs/xfs/libxfs/xfs_inode_buf.c
+@@ -29,6 +29,7 @@
+ #include "xfs_icache.h"
+ #include "xfs_trans.h"
+ #include "xfs_ialloc.h"
++#include "xfs_dir2.h"
+ 
+ /*
+  * Check that none of the inode's in the buffer have a next
+@@ -386,6 +387,7 @@ xfs_dinode_verify(
+       struct xfs_inode        *ip,
+       struct xfs_dinode       *dip)
+ {
++      uint16_t                mode;
+       uint16_t                flags;
+       uint64_t                flags2;
+ 
+@@ -396,8 +398,12 @@ xfs_dinode_verify(
+       if (be64_to_cpu(dip->di_size) & (1ULL << 63))
+               return false;
+ 
+-      /* No zero-length symlinks. */
+-      if (S_ISLNK(be16_to_cpu(dip->di_mode)) && dip->di_size == 0)
++      mode = be16_to_cpu(dip->di_mode);
++      if (mode && xfs_mode_to_ftype(mode) == XFS_DIR3_FT_UNKNOWN)
++              return false;
++
++      /* No zero-length symlinks/dirs. */
++      if ((S_ISLNK(mode) || S_ISDIR(mode)) && dip->di_size == 0)
+               return false;
+ 
+       /* only version 3 or greater inodes are extensively verified here */
+diff --git a/fs/xfs/libxfs/xfs_sb.c b/fs/xfs/libxfs/xfs_sb.c
+index 2580262e4ea0..584ec896a533 100644
+--- a/fs/xfs/libxfs/xfs_sb.c
++++ b/fs/xfs/libxfs/xfs_sb.c
+@@ -242,7 +242,7 @@ xfs_mount_validate_sb(
+           sbp->sb_blocklog < XFS_MIN_BLOCKSIZE_LOG                    ||
+           sbp->sb_blocklog > XFS_MAX_BLOCKSIZE_LOG                    ||
+           sbp->sb_blocksize != (1 << sbp->sb_blocklog)                ||
+-          sbp->sb_dirblklog > XFS_MAX_BLOCKSIZE_LOG                   ||
++          sbp->sb_dirblklog + sbp->sb_blocklog > XFS_MAX_BLOCKSIZE_LOG ||
+           sbp->sb_inodesize < XFS_DINODE_MIN_SIZE                     ||
+           sbp->sb_inodesize > XFS_DINODE_MAX_SIZE                     ||
+           sbp->sb_inodelog < XFS_DINODE_MIN_LOG                       ||
+diff --git a/fs/xfs/xfs_bmap_util.c b/fs/xfs/xfs_bmap_util.c
+index 0670a8bd5818..efb8ccd6bbf2 100644
+--- a/fs/xfs/xfs_bmap_util.c
++++ b/fs/xfs/xfs_bmap_util.c
+@@ -528,7 +528,6 @@ xfs_getbmap(
+       xfs_bmbt_irec_t         *map;           /* buffer for user's data */
+       xfs_mount_t             *mp;            /* file system mount point */
+       int                     nex;            /* # of user extents can do */
+-      int                     nexleft;        /* # of user extents left */
+       int                     subnex;         /* # of bmapi's can do */
+       int                     nmap;           /* number of map entries */
+       struct getbmapx         *out;           /* output structure */
+@@ -686,10 +685,8 @@ xfs_getbmap(
+               goto out_free_map;
+       }
+ 
+-      nexleft = nex;
+-
+       do {
+-              nmap = (nexleft > subnex) ? subnex : nexleft;
++              nmap = (nex> subnex) ? subnex : nex;
+               error = xfs_bmapi_read(ip, XFS_BB_TO_FSBT(mp, bmv->bmv_offset),
+                                      XFS_BB_TO_FSB(mp, bmv->bmv_length),
+                                      map, &nmap, bmapi_flags);
+@@ -697,8 +694,8 @@ xfs_getbmap(
+                       goto out_free_map;
+               ASSERT(nmap <= subnex);
+ 
+-              for (i = 0; i < nmap && nexleft && bmv->bmv_length &&
+-                              cur_ext < bmv->bmv_count; i++) {
++              for (i = 0; i < nmap && bmv->bmv_length &&
++                              cur_ext < bmv->bmv_count - 1; i++) {
+                       out[cur_ext].bmv_oflags = 0;
+                       if (map[i].br_state == XFS_EXT_UNWRITTEN)
+                               out[cur_ext].bmv_oflags |= BMV_OF_PREALLOC;
+@@ -760,16 +757,27 @@ xfs_getbmap(
+                               continue;
+                       }
+ 
++                      /*
++                       * In order to report shared extents accurately,
++                       * we report each distinct shared/unshared part
++                       * of a single bmbt record using multiple bmap
++                       * extents.  To make that happen, we iterate the
++                       * same map array item multiple times, each
++                       * time trimming out the subextent that we just
++                       * reported.
++                       *
++                       * Because of this, we must check the out array
++                       * index (cur_ext) directly against bmv_count-1
++                       * to avoid overflows.
++                       */
+                       if (inject_map.br_startblock != NULLFSBLOCK) {
+                               map[i] = inject_map;
+                               i--;
+-                      } else
+-                              nexleft--;
++                      }
+                       bmv->bmv_entries++;
+                       cur_ext++;
+               }
+-      } while (nmap && nexleft && bmv->bmv_length &&
+-               cur_ext < bmv->bmv_count);
++      } while (nmap && bmv->bmv_length && cur_ext < bmv->bmv_count - 1);
+ 
+  out_free_map:
+       kmem_free(map);
+diff --git a/fs/xfs/xfs_buf.c b/fs/xfs/xfs_buf.c
+index b5b9bffe3520..d7a67d7fbc7f 100644
+--- a/fs/xfs/xfs_buf.c
++++ b/fs/xfs/xfs_buf.c
+@@ -423,6 +423,7 @@ xfs_buf_allocate_memory(
+ out_free_pages:
+       for (i = 0; i < bp->b_page_count; i++)
+               __free_page(bp->b_pages[i]);
++      bp->b_flags &= ~_XBF_PAGES;
+       return error;
+ }
+ 
+diff --git a/fs/xfs/xfs_dquot.c b/fs/xfs/xfs_dquot.c
+index 7a30b8f11db7..9d06cc30e875 100644
+--- a/fs/xfs/xfs_dquot.c
++++ b/fs/xfs/xfs_dquot.c
+@@ -710,6 +710,10 @@ xfs_dq_get_next_id(
+       /* Simple advance */
+       next_id = *id + 1;
+ 
++      /* If we'd wrap past the max ID, stop */
++      if (next_id < *id)
++              return -ENOENT;
++
+       /* If new ID is within the current chunk, advancing it sufficed */
+       if (next_id % mp->m_quotainfo->qi_dqperchunk) {
+               *id = next_id;
+diff --git a/fs/xfs/xfs_iomap.c b/fs/xfs/xfs_iomap.c
+index 15a83813b708..cdc6bdd495be 100644
+--- a/fs/xfs/xfs_iomap.c
++++ b/fs/xfs/xfs_iomap.c
+@@ -681,7 +681,7 @@ xfs_iomap_write_allocate(
+       xfs_trans_t     *tp;
+       int             nimaps;
+       int             error = 0;
+-      int             flags = 0;
++      int             flags = XFS_BMAPI_DELALLOC;
+       int             nres;
+ 
+       if (whichfork == XFS_COW_FORK)
+diff --git a/fs/xfs/xfs_iops.c b/fs/xfs/xfs_iops.c
+index 405a65cd9d6b..f5e0f608e245 100644
+--- a/fs/xfs/xfs_iops.c
++++ b/fs/xfs/xfs_iops.c
+@@ -98,12 +98,27 @@ xfs_init_security(
+ static void
+ xfs_dentry_to_name(
+       struct xfs_name *namep,
++      struct dentry   *dentry)
++{
++      namep->name = dentry->d_name.name;
++      namep->len = dentry->d_name.len;
++      namep->type = XFS_DIR3_FT_UNKNOWN;
++}
++
++static int
++xfs_dentry_mode_to_name(
++      struct xfs_name *namep,
+       struct dentry   *dentry,
+       int             mode)
+ {
+       namep->name = dentry->d_name.name;
+       namep->len = dentry->d_name.len;
+-      namep->type = xfs_mode_to_ftype[(mode & S_IFMT) >> S_SHIFT];
++      namep->type = xfs_mode_to_ftype(mode);
++
++      if (unlikely(namep->type == XFS_DIR3_FT_UNKNOWN))
++              return -EFSCORRUPTED;
++
++      return 0;
+ }
+ 
+ STATIC void
+@@ -119,7 +134,7 @@ xfs_cleanup_inode(
+        * xfs_init_security we must back out.
+        * ENOSPC can hit here, among other things.
+        */
+-      xfs_dentry_to_name(&teardown, dentry, 0);
++      xfs_dentry_to_name(&teardown, dentry);
+ 
+       xfs_remove(XFS_I(dir), &teardown, XFS_I(inode));
+ }
+@@ -154,8 +169,12 @@ xfs_generic_create(
+       if (error)
+               return error;
+ 
++      /* Verify mode is valid also for tmpfile case */
++      error = xfs_dentry_mode_to_name(&name, dentry, mode);
++      if (unlikely(error))
++              goto out_free_acl;
++
+       if (!tmpfile) {
+-              xfs_dentry_to_name(&name, dentry, mode);
+               error = xfs_create(XFS_I(dir), &name, mode, rdev, &ip);
+       } else {
+               error = xfs_create_tmpfile(XFS_I(dir), dentry, mode, &ip);
+@@ -248,7 +267,7 @@ xfs_vn_lookup(
+       if (dentry->d_name.len >= MAXNAMELEN)
+               return ERR_PTR(-ENAMETOOLONG);
+ 
+-      xfs_dentry_to_name(&name, dentry, 0);
++      xfs_dentry_to_name(&name, dentry);
+       error = xfs_lookup(XFS_I(dir), &name, &cip, NULL);
+       if (unlikely(error)) {
+               if (unlikely(error != -ENOENT))
+@@ -275,7 +294,7 @@ xfs_vn_ci_lookup(
+       if (dentry->d_name.len >= MAXNAMELEN)
+               return ERR_PTR(-ENAMETOOLONG);
+ 
+-      xfs_dentry_to_name(&xname, dentry, 0);
++      xfs_dentry_to_name(&xname, dentry);
+       error = xfs_lookup(XFS_I(dir), &xname, &ip, &ci_name);
+       if (unlikely(error)) {
+               if (unlikely(error != -ENOENT))
+@@ -310,7 +329,9 @@ xfs_vn_link(
+       struct xfs_name name;
+       int             error;
+ 
+-      xfs_dentry_to_name(&name, dentry, inode->i_mode);
++      error = xfs_dentry_mode_to_name(&name, dentry, inode->i_mode);
++      if (unlikely(error))
++              return error;
+ 
+       error = xfs_link(XFS_I(dir), XFS_I(inode), &name);
+       if (unlikely(error))
+@@ -329,7 +350,7 @@ xfs_vn_unlink(
+       struct xfs_name name;
+       int             error;
+ 
+-      xfs_dentry_to_name(&name, dentry, 0);
++      xfs_dentry_to_name(&name, dentry);
+ 
+       error = xfs_remove(XFS_I(dir), &name, XFS_I(d_inode(dentry)));
+       if (error)
+@@ -359,7 +380,9 @@ xfs_vn_symlink(
+ 
+       mode = S_IFLNK |
+               (irix_symlink_mode ? 0777 & ~current_umask() : S_IRWXUGO);
+-      xfs_dentry_to_name(&name, dentry, mode);
++      error = xfs_dentry_mode_to_name(&name, dentry, mode);
++      if (unlikely(error))
++              goto out;
+ 
+       error = xfs_symlink(XFS_I(dir), &name, symname, mode, &cip);
+       if (unlikely(error))
+@@ -395,6 +418,7 @@ xfs_vn_rename(
+ {
+       struct inode    *new_inode = d_inode(ndentry);
+       int             omode = 0;
++      int             error;
+       struct xfs_name oname;
+       struct xfs_name nname;
+ 
+@@ -405,8 +429,14 @@ xfs_vn_rename(
+       if (flags & RENAME_EXCHANGE)
+               omode = d_inode(ndentry)->i_mode;
+ 
+-      xfs_dentry_to_name(&oname, odentry, omode);
+-      xfs_dentry_to_name(&nname, ndentry, d_inode(odentry)->i_mode);
++      error = xfs_dentry_mode_to_name(&oname, odentry, omode);
++      if (omode && unlikely(error))
++              return error;
++
++      error = xfs_dentry_mode_to_name(&nname, ndentry,
++                                      d_inode(odentry)->i_mode);
++      if (unlikely(error))
++              return error;
+ 
+       return xfs_rename(XFS_I(odir), &oname, XFS_I(d_inode(odentry)),
+                         XFS_I(ndir), &nname,
+diff --git a/fs/xfs/xfs_linux.h b/fs/xfs/xfs_linux.h
+index 68640fb63a54..1455b25205a8 100644
+--- a/fs/xfs/xfs_linux.h
++++ b/fs/xfs/xfs_linux.h
+@@ -330,11 +330,11 @@ static inline __uint64_t howmany_64(__uint64_t x, 
__uint32_t y)
+ }
+ 
+ #define ASSERT_ALWAYS(expr)   \
+-      (unlikely(expr) ? (void)0 : assfail(#expr, __FILE__, __LINE__))
++      (likely(expr) ? (void)0 : assfail(#expr, __FILE__, __LINE__))
+ 
+ #ifdef DEBUG
+ #define ASSERT(expr)  \
+-      (unlikely(expr) ? (void)0 : assfail(#expr, __FILE__, __LINE__))
++      (likely(expr) ? (void)0 : assfail(#expr, __FILE__, __LINE__))
+ 
+ #ifndef STATIC
+ # define STATIC noinline
+@@ -345,7 +345,7 @@ static inline __uint64_t howmany_64(__uint64_t x, 
__uint32_t y)
+ #ifdef XFS_WARN
+ 
+ #define ASSERT(expr)  \
+-      (unlikely(expr) ? (void)0 : asswarn(#expr, __FILE__, __LINE__))
++      (likely(expr) ? (void)0 : asswarn(#expr, __FILE__, __LINE__))
+ 
+ #ifndef STATIC
+ # define STATIC static noinline
+diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c
+index 3b74fa011bb1..4017aa967331 100644
+--- a/fs/xfs/xfs_log.c
++++ b/fs/xfs/xfs_log.c
+@@ -3324,12 +3324,8 @@ xfs_log_force(
+       xfs_mount_t     *mp,
+       uint            flags)
+ {
+-      int     error;
+-
+       trace_xfs_log_force(mp, 0, _RET_IP_);
+-      error = _xfs_log_force(mp, flags, NULL);
+-      if (error)
+-              xfs_warn(mp, "%s: error %d returned.", __func__, error);
++      _xfs_log_force(mp, flags, NULL);
+ }
+ 
+ /*
+@@ -3473,12 +3469,8 @@ xfs_log_force_lsn(
+       xfs_lsn_t       lsn,
+       uint            flags)
+ {
+-      int     error;
+-
+       trace_xfs_log_force(mp, lsn, _RET_IP_);
+-      error = _xfs_log_force_lsn(mp, lsn, flags, NULL);
+-      if (error)
+-              xfs_warn(mp, "%s: error %d returned.", __func__, error);
++      _xfs_log_force_lsn(mp, lsn, flags, NULL);
+ }
+ 
+ /*
+diff --git a/include/linux/tcp.h b/include/linux/tcp.h
+index a17ae7b85218..647532b0eb03 100644
+--- a/include/linux/tcp.h
++++ b/include/linux/tcp.h
+@@ -62,8 +62,13 @@ static inline unsigned int tcp_optlen(const struct sk_buff 
*skb)
+ 
+ /* TCP Fast Open Cookie as stored in memory */
+ struct tcp_fastopen_cookie {
++      union {
++              u8      val[TCP_FASTOPEN_COOKIE_MAX];
++#if IS_ENABLED(CONFIG_IPV6)
++              struct in6_addr addr;
++#endif
++      };
+       s8      len;
+-      u8      val[TCP_FASTOPEN_COOKIE_MAX];
+       bool    exp;    /* In RFC6994 experimental option format */
+ };
+ 
+diff --git a/include/linux/virtio_net.h b/include/linux/virtio_net.h
+index 1c912f85e041..f211c348e592 100644
+--- a/include/linux/virtio_net.h
++++ b/include/linux/virtio_net.h
+@@ -56,7 +56,8 @@ static inline int virtio_net_hdr_to_skb(struct sk_buff *skb,
+ 
+ static inline int virtio_net_hdr_from_skb(const struct sk_buff *skb,
+                                         struct virtio_net_hdr *hdr,
+-                                        bool little_endian)
++                                        bool little_endian,
++                                        bool has_data_valid)
+ {
+       memset(hdr, 0, sizeof(*hdr));
+ 
+@@ -91,7 +92,8 @@ static inline int virtio_net_hdr_from_skb(const struct 
sk_buff *skb,
+                               skb_checksum_start_offset(skb));
+               hdr->csum_offset = __cpu_to_virtio16(little_endian,
+                               skb->csum_offset);
+-      } else if (skb->ip_summed == CHECKSUM_UNNECESSARY) {
++      } else if (has_data_valid &&
++                 skb->ip_summed == CHECKSUM_UNNECESSARY) {
+               hdr->flags = VIRTIO_NET_HDR_F_DATA_VALID;
+       } /* else everything is zero */
+ 
+diff --git a/include/net/lwtunnel.h b/include/net/lwtunnel.h
+index ea3f80f58fd6..fc7c0dbdd1ff 100644
+--- a/include/net/lwtunnel.h
++++ b/include/net/lwtunnel.h
+@@ -43,13 +43,12 @@ struct lwtunnel_encap_ops {
+       int (*get_encap_size)(struct lwtunnel_state *lwtstate);
+       int (*cmp_encap)(struct lwtunnel_state *a, struct lwtunnel_state *b);
+       int (*xmit)(struct sk_buff *skb);
++
++      struct module *owner;
+ };
+ 
+ #ifdef CONFIG_LWTUNNEL
+-static inline void lwtstate_free(struct lwtunnel_state *lws)
+-{
+-      kfree(lws);
+-}
++void lwtstate_free(struct lwtunnel_state *lws);
+ 
+ static inline struct lwtunnel_state *
+ lwtstate_get(struct lwtunnel_state *lws)
+@@ -106,6 +105,8 @@ int lwtunnel_encap_add_ops(const struct lwtunnel_encap_ops 
*op,
+                          unsigned int num);
+ int lwtunnel_encap_del_ops(const struct lwtunnel_encap_ops *op,
+                          unsigned int num);
++int lwtunnel_valid_encap_type(u16 encap_type);
++int lwtunnel_valid_encap_type_attr(struct nlattr *attr, int len);
+ int lwtunnel_build_state(struct net_device *dev, u16 encap_type,
+                        struct nlattr *encap,
+                        unsigned int family, const void *cfg,
+@@ -169,6 +170,15 @@ static inline int lwtunnel_encap_del_ops(const struct 
lwtunnel_encap_ops *op,
+       return -EOPNOTSUPP;
+ }
+ 
++static inline int lwtunnel_valid_encap_type(u16 encap_type)
++{
++      return -EOPNOTSUPP;
++}
++static inline int lwtunnel_valid_encap_type_attr(struct nlattr *attr, int len)
++{
++      return -EOPNOTSUPP;
++}
++
+ static inline int lwtunnel_build_state(struct net_device *dev, u16 encap_type,
+                                      struct nlattr *encap,
+                                      unsigned int family, const void *cfg,
+diff --git a/net/ax25/ax25_subr.c b/net/ax25/ax25_subr.c
+index 655a7d4c96e1..983f0b5e14f1 100644
+--- a/net/ax25/ax25_subr.c
++++ b/net/ax25/ax25_subr.c
+@@ -264,7 +264,7 @@ void ax25_disconnect(ax25_cb *ax25, int reason)
+ {
+       ax25_clear_queues(ax25);
+ 
+-      if (!sock_flag(ax25->sk, SOCK_DESTROY))
++      if (!ax25->sk || !sock_flag(ax25->sk, SOCK_DESTROY))
+               ax25_stop_heartbeat(ax25);
+       ax25_stop_t1timer(ax25);
+       ax25_stop_t2timer(ax25);
+diff --git a/net/bridge/br_netlink.c b/net/bridge/br_netlink.c
+index e99037c6f7b7..04741064a173 100644
+--- a/net/bridge/br_netlink.c
++++ b/net/bridge/br_netlink.c
+@@ -781,20 +781,6 @@ static int br_validate(struct nlattr *tb[], struct nlattr 
*data[])
+       return 0;
+ }
+ 
+-static int br_dev_newlink(struct net *src_net, struct net_device *dev,
+-                        struct nlattr *tb[], struct nlattr *data[])
+-{
+-      struct net_bridge *br = netdev_priv(dev);
+-
+-      if (tb[IFLA_ADDRESS]) {
+-              spin_lock_bh(&br->lock);
+-              br_stp_change_bridge_id(br, nla_data(tb[IFLA_ADDRESS]));
+-              spin_unlock_bh(&br->lock);
+-      }
+-
+-      return register_netdevice(dev);
+-}
+-
+ static int br_port_slave_changelink(struct net_device *brdev,
+                                   struct net_device *dev,
+                                   struct nlattr *tb[],
+@@ -1093,6 +1079,25 @@ static int br_changelink(struct net_device *brdev, 
struct nlattr *tb[],
+       return 0;
+ }
+ 
++static int br_dev_newlink(struct net *src_net, struct net_device *dev,
++                        struct nlattr *tb[], struct nlattr *data[])
++{
++      struct net_bridge *br = netdev_priv(dev);
++      int err;
++
++      if (tb[IFLA_ADDRESS]) {
++              spin_lock_bh(&br->lock);
++              br_stp_change_bridge_id(br, nla_data(tb[IFLA_ADDRESS]));
++              spin_unlock_bh(&br->lock);
++      }
++
++      err = br_changelink(dev, tb, data);
++      if (err)
++              return err;
++
++      return register_netdevice(dev);
++}
++
+ static size_t br_get_size(const struct net_device *brdev)
+ {
+       return nla_total_size(sizeof(u32)) +    /* IFLA_BR_FORWARD_DELAY  */
+diff --git a/net/core/dev.c b/net/core/dev.c
+index e1d731fdc72c..df51c50927ab 100644
+--- a/net/core/dev.c
++++ b/net/core/dev.c
+@@ -2815,9 +2815,9 @@ static netdev_features_t harmonize_features(struct 
sk_buff *skb,
+       if (skb->ip_summed != CHECKSUM_NONE &&
+           !can_checksum_protocol(features, type)) {
+               features &= ~(NETIF_F_CSUM_MASK | NETIF_F_GSO_MASK);
+-      } else if (illegal_highdma(skb->dev, skb)) {
+-              features &= ~NETIF_F_SG;
+       }
++      if (illegal_highdma(skb->dev, skb))
++              features &= ~NETIF_F_SG;
+ 
+       return features;
+ }
+diff --git a/net/core/lwtunnel.c b/net/core/lwtunnel.c
+index e5f84c26ba1a..afa64f086d87 100644
+--- a/net/core/lwtunnel.c
++++ b/net/core/lwtunnel.c
+@@ -26,6 +26,7 @@
+ #include <net/lwtunnel.h>
+ #include <net/rtnetlink.h>
+ #include <net/ip6_fib.h>
++#include <net/nexthop.h>
+ 
+ #ifdef CONFIG_MODULES
+ 
+@@ -65,6 +66,15 @@ EXPORT_SYMBOL(lwtunnel_state_alloc);
+ static const struct lwtunnel_encap_ops __rcu *
+               lwtun_encaps[LWTUNNEL_ENCAP_MAX + 1] __read_mostly;
+ 
++void lwtstate_free(struct lwtunnel_state *lws)
++{
++      const struct lwtunnel_encap_ops *ops = lwtun_encaps[lws->type];
++
++      kfree(lws);
++      module_put(ops->owner);
++}
++EXPORT_SYMBOL(lwtstate_free);
++
+ int lwtunnel_encap_add_ops(const struct lwtunnel_encap_ops *ops,
+                          unsigned int num)
+ {
+@@ -110,25 +120,77 @@ int lwtunnel_build_state(struct net_device *dev, u16 
encap_type,
+       ret = -EOPNOTSUPP;
+       rcu_read_lock();
+       ops = rcu_dereference(lwtun_encaps[encap_type]);
++      if (likely(ops && ops->build_state && try_module_get(ops->owner))) {
++              ret = ops->build_state(dev, encap, family, cfg, lws);
++              if (ret)
++                      module_put(ops->owner);
++      }
++      rcu_read_unlock();
++
++      return ret;
++}
++EXPORT_SYMBOL(lwtunnel_build_state);
++
++int lwtunnel_valid_encap_type(u16 encap_type)
++{
++      const struct lwtunnel_encap_ops *ops;
++      int ret = -EINVAL;
++
++      if (encap_type == LWTUNNEL_ENCAP_NONE ||
++          encap_type > LWTUNNEL_ENCAP_MAX)
++              return ret;
++
++      rcu_read_lock();
++      ops = rcu_dereference(lwtun_encaps[encap_type]);
++      rcu_read_unlock();
+ #ifdef CONFIG_MODULES
+       if (!ops) {
+               const char *encap_type_str = lwtunnel_encap_str(encap_type);
+ 
+               if (encap_type_str) {
+-                      rcu_read_unlock();
++                      __rtnl_unlock();
+                       request_module("rtnl-lwt-%s", encap_type_str);
++                      rtnl_lock();
++
+                       rcu_read_lock();
+                       ops = rcu_dereference(lwtun_encaps[encap_type]);
++                      rcu_read_unlock();
+               }
+       }
+ #endif
+-      if (likely(ops && ops->build_state))
+-              ret = ops->build_state(dev, encap, family, cfg, lws);
+-      rcu_read_unlock();
++      return ops ? 0 : -EOPNOTSUPP;
++}
++EXPORT_SYMBOL(lwtunnel_valid_encap_type);
+ 
+-      return ret;
++int lwtunnel_valid_encap_type_attr(struct nlattr *attr, int remaining)
++{
++      struct rtnexthop *rtnh = (struct rtnexthop *)attr;
++      struct nlattr *nla_entype;
++      struct nlattr *attrs;
++      struct nlattr *nla;
++      u16 encap_type;
++      int attrlen;
++
++      while (rtnh_ok(rtnh, remaining)) {
++              attrlen = rtnh_attrlen(rtnh);
++              if (attrlen > 0) {
++                      attrs = rtnh_attrs(rtnh);
++                      nla = nla_find(attrs, attrlen, RTA_ENCAP);
++                      nla_entype = nla_find(attrs, attrlen, RTA_ENCAP_TYPE);
++
++                      if (nla_entype) {
++                              encap_type = nla_get_u16(nla_entype);
++
++                              if (lwtunnel_valid_encap_type(encap_type) != 0)
++                                      return -EOPNOTSUPP;
++                      }
++              }
++              rtnh = rtnh_next(rtnh, &remaining);
++      }
++
++      return 0;
+ }
+-EXPORT_SYMBOL(lwtunnel_build_state);
++EXPORT_SYMBOL(lwtunnel_valid_encap_type_attr);
+ 
+ int lwtunnel_fill_encap(struct sk_buff *skb, struct lwtunnel_state *lwtstate)
+ {
+diff --git a/net/dsa/slave.c b/net/dsa/slave.c
+index 30e2e21d7619..3ff9d97cf56b 100644
+--- a/net/dsa/slave.c
++++ b/net/dsa/slave.c
+@@ -1201,6 +1201,8 @@ int dsa_slave_suspend(struct net_device *slave_dev)
+ {
+       struct dsa_slave_priv *p = netdev_priv(slave_dev);
+ 
++      netif_device_detach(slave_dev);
++
+       if (p->phy) {
+               phy_stop(p->phy);
+               p->old_pause = -1;
+diff --git a/net/ipv4/fib_frontend.c b/net/ipv4/fib_frontend.c
+index 3e4f183fc241..5b03d7f3b255 100644
+--- a/net/ipv4/fib_frontend.c
++++ b/net/ipv4/fib_frontend.c
+@@ -46,6 +46,7 @@
+ #include <net/rtnetlink.h>
+ #include <net/xfrm.h>
+ #include <net/l3mdev.h>
++#include <net/lwtunnel.h>
+ #include <trace/events/fib.h>
+ 
+ #ifndef CONFIG_IP_MULTIPLE_TABLES
+@@ -676,6 +677,10 @@ static int rtm_to_fib_config(struct net *net, struct 
sk_buff *skb,
+                       cfg->fc_mx_len = nla_len(attr);
+                       break;
+               case RTA_MULTIPATH:
++                      err = lwtunnel_valid_encap_type_attr(nla_data(attr),
++                                                           nla_len(attr));
++                      if (err < 0)
++                              goto errout;
+                       cfg->fc_mp = nla_data(attr);
+                       cfg->fc_mp_len = nla_len(attr);
+                       break;
+@@ -690,6 +695,9 @@ static int rtm_to_fib_config(struct net *net, struct 
sk_buff *skb,
+                       break;
+               case RTA_ENCAP_TYPE:
+                       cfg->fc_encap_type = nla_get_u16(attr);
++                      err = lwtunnel_valid_encap_type(cfg->fc_encap_type);
++                      if (err < 0)
++                              goto errout;
+                       break;
+               }
+       }
+diff --git a/net/ipv4/fib_semantics.c b/net/ipv4/fib_semantics.c
+index a8508b79b406..6a4068031aaa 100644
+--- a/net/ipv4/fib_semantics.c
++++ b/net/ipv4/fib_semantics.c
+@@ -1278,8 +1278,9 @@ int fib_dump_info(struct sk_buff *skb, u32 portid, u32 
seq, int event,
+                   nla_put_u32(skb, RTA_FLOW, fi->fib_nh[0].nh_tclassid))
+                       goto nla_put_failure;
+ #endif
+-              if (fi->fib_nh->nh_lwtstate)
+-                      lwtunnel_fill_encap(skb, fi->fib_nh->nh_lwtstate);
++              if (fi->fib_nh->nh_lwtstate &&
++                  lwtunnel_fill_encap(skb, fi->fib_nh->nh_lwtstate) < 0)
++                      goto nla_put_failure;
+       }
+ #ifdef CONFIG_IP_ROUTE_MULTIPATH
+       if (fi->fib_nhs > 1) {
+@@ -1315,8 +1316,10 @@ int fib_dump_info(struct sk_buff *skb, u32 portid, u32 
seq, int event,
+                           nla_put_u32(skb, RTA_FLOW, nh->nh_tclassid))
+                               goto nla_put_failure;
+ #endif
+-                      if (nh->nh_lwtstate)
+-                              lwtunnel_fill_encap(skb, nh->nh_lwtstate);
++                      if (nh->nh_lwtstate &&
++                          lwtunnel_fill_encap(skb, nh->nh_lwtstate) < 0)
++                              goto nla_put_failure;
++
+                       /* length of rtnetlink header + attributes */
+                       rtnh->rtnh_len = nlmsg_get_pos(skb) - (void *) rtnh;
+               } endfor_nexthops(fi);
+diff --git a/net/ipv4/ip_tunnel_core.c b/net/ipv4/ip_tunnel_core.c
+index fed3d29f9eb3..0fd1976ab63b 100644
+--- a/net/ipv4/ip_tunnel_core.c
++++ b/net/ipv4/ip_tunnel_core.c
+@@ -313,6 +313,7 @@ static const struct lwtunnel_encap_ops ip_tun_lwt_ops = {
+       .fill_encap = ip_tun_fill_encap_info,
+       .get_encap_size = ip_tun_encap_nlsize,
+       .cmp_encap = ip_tun_cmp_encap,
++      .owner = THIS_MODULE,
+ };
+ 
+ static const struct nla_policy ip6_tun_policy[LWTUNNEL_IP6_MAX + 1] = {
+@@ -403,6 +404,7 @@ static const struct lwtunnel_encap_ops ip6_tun_lwt_ops = {
+       .fill_encap = ip6_tun_fill_encap_info,
+       .get_encap_size = ip6_tun_encap_nlsize,
+       .cmp_encap = ip_tun_cmp_encap,
++      .owner = THIS_MODULE,
+ };
+ 
+ void __init ip_tunnel_core_init(void)
+diff --git a/net/ipv4/route.c b/net/ipv4/route.c
+index 8197b06d9aaa..d851cae27dac 100644
+--- a/net/ipv4/route.c
++++ b/net/ipv4/route.c
+@@ -2440,7 +2440,7 @@ static int rt_fill_info(struct net *net,  __be32 dst, 
__be32 src, u32 table_id,
+       r->rtm_dst_len  = 32;
+       r->rtm_src_len  = 0;
+       r->rtm_tos      = fl4->flowi4_tos;
+-      r->rtm_table    = table_id;
++      r->rtm_table    = table_id < 256 ? table_id : RT_TABLE_COMPAT;
+       if (nla_put_u32(skb, RTA_TABLE, table_id))
+               goto nla_put_failure;
+       r->rtm_type     = rt->rt_type;
+diff --git a/net/ipv4/tcp_fastopen.c b/net/ipv4/tcp_fastopen.c
+index 4e777a3243f9..dd2560c83a85 100644
+--- a/net/ipv4/tcp_fastopen.c
++++ b/net/ipv4/tcp_fastopen.c
+@@ -113,7 +113,7 @@ static bool tcp_fastopen_cookie_gen(struct request_sock 
*req,
+               struct tcp_fastopen_cookie tmp;
+ 
+               if (__tcp_fastopen_cookie_gen(&ip6h->saddr, &tmp)) {
+-                      struct in6_addr *buf = (struct in6_addr *) tmp.val;
++                      struct in6_addr *buf = &tmp.addr;
+                       int i;
+ 
+                       for (i = 0; i < 4; i++)
+@@ -205,6 +205,7 @@ static struct sock *tcp_fastopen_create_child(struct sock 
*sk,
+        * scaled. So correct it appropriately.
+        */
+       tp->snd_wnd = ntohs(tcp_hdr(skb)->window);
++      tp->max_window = tp->snd_wnd;
+ 
+       /* Activate the retrans timer so that SYNACK can be retransmitted.
+        * The request socket is not added to the ehash
+diff --git a/net/ipv6/addrconf.c b/net/ipv6/addrconf.c
+index 4bc5ba3ae452..95dfcba38ff6 100644
+--- a/net/ipv6/addrconf.c
++++ b/net/ipv6/addrconf.c
+@@ -5515,8 +5515,7 @@ static void addrconf_disable_change(struct net *net, 
__s32 newf)
+       struct net_device *dev;
+       struct inet6_dev *idev;
+ 
+-      rcu_read_lock();
+-      for_each_netdev_rcu(net, dev) {
++      for_each_netdev(net, dev) {
+               idev = __in6_dev_get(dev);
+               if (idev) {
+                       int changed = (!idev->cnf.disable_ipv6) ^ (!newf);
+@@ -5525,7 +5524,6 @@ static void addrconf_disable_change(struct net *net, 
__s32 newf)
+                               dev_disable_change(idev);
+               }
+       }
+-      rcu_read_unlock();
+ }
+ 
+ static int addrconf_disable_ipv6(struct ctl_table *table, int *p, int newf)
+diff --git a/net/ipv6/ila/ila_lwt.c b/net/ipv6/ila/ila_lwt.c
+index e50c27a93e17..f3db364fc853 100644
+--- a/net/ipv6/ila/ila_lwt.c
++++ b/net/ipv6/ila/ila_lwt.c
+@@ -164,6 +164,7 @@ static const struct lwtunnel_encap_ops ila_encap_ops = {
+       .fill_encap = ila_fill_encap_info,
+       .get_encap_size = ila_encap_nlsize,
+       .cmp_encap = ila_encap_cmp,
++      .owner = THIS_MODULE,
+ };
+ 
+ int ila_lwt_init(void)
+diff --git a/net/ipv6/ip6_tunnel.c b/net/ipv6/ip6_tunnel.c
+index d76674efe523..f95437f1087c 100644
+--- a/net/ipv6/ip6_tunnel.c
++++ b/net/ipv6/ip6_tunnel.c
+@@ -1108,7 +1108,7 @@ int ip6_tnl_xmit(struct sk_buff *skb, struct net_device 
*dev, __u8 dsfield,
+                                    t->parms.name);
+               goto tx_err_dst_release;
+       }
+-      mtu = dst_mtu(dst) - psh_hlen;
++      mtu = dst_mtu(dst) - psh_hlen - t->tun_hlen;
+       if (encap_limit >= 0) {
+               max_headroom += 8;
+               mtu -= 8;
+@@ -1117,7 +1117,7 @@ int ip6_tnl_xmit(struct sk_buff *skb, struct net_device 
*dev, __u8 dsfield,
+               mtu = IPV6_MIN_MTU;
+       if (skb_dst(skb) && !t->parms.collect_md)
+               skb_dst(skb)->ops->update_pmtu(skb_dst(skb), NULL, skb, mtu);
+-      if (skb->len > mtu && !skb_is_gso(skb)) {
++      if (skb->len - t->tun_hlen > mtu && !skb_is_gso(skb)) {
+               *pmtu = mtu;
+               err = -EMSGSIZE;
+               goto tx_err_dst_release;
+diff --git a/net/ipv6/route.c b/net/ipv6/route.c
+index 1b57e11e6e0d..bff4460f17be 100644
+--- a/net/ipv6/route.c
++++ b/net/ipv6/route.c
+@@ -2885,6 +2885,11 @@ static int rtm_to_fib6_config(struct sk_buff *skb, 
struct nlmsghdr *nlh,
+       if (tb[RTA_MULTIPATH]) {
+               cfg->fc_mp = nla_data(tb[RTA_MULTIPATH]);
+               cfg->fc_mp_len = nla_len(tb[RTA_MULTIPATH]);
++
++              err = lwtunnel_valid_encap_type_attr(cfg->fc_mp,
++                                                   cfg->fc_mp_len);
++              if (err < 0)
++                      goto errout;
+       }
+ 
+       if (tb[RTA_PREF]) {
+@@ -2898,9 +2903,14 @@ static int rtm_to_fib6_config(struct sk_buff *skb, 
struct nlmsghdr *nlh,
+       if (tb[RTA_ENCAP])
+               cfg->fc_encap = tb[RTA_ENCAP];
+ 
+-      if (tb[RTA_ENCAP_TYPE])
++      if (tb[RTA_ENCAP_TYPE]) {
+               cfg->fc_encap_type = nla_get_u16(tb[RTA_ENCAP_TYPE]);
+ 
++              err = lwtunnel_valid_encap_type(cfg->fc_encap_type);
++              if (err < 0)
++                      goto errout;
++      }
++
+       if (tb[RTA_EXPIRES]) {
+               unsigned long timeout = 
addrconf_timeout_fixup(nla_get_u32(tb[RTA_EXPIRES]), HZ);
+ 
+@@ -3306,7 +3316,8 @@ static int rt6_fill_node(struct net *net,
+       if (nla_put_u8(skb, RTA_PREF, IPV6_EXTRACT_PREF(rt->rt6i_flags)))
+               goto nla_put_failure;
+ 
+-      lwtunnel_fill_encap(skb, rt->dst.lwtstate);
++      if (lwtunnel_fill_encap(skb, rt->dst.lwtstate) < 0)
++              goto nla_put_failure;
+ 
+       nlmsg_end(skb, nlh);
+       return 0;
+diff --git a/net/mpls/af_mpls.c b/net/mpls/af_mpls.c
+index 15fe97644ffe..5b77377e5a15 100644
+--- a/net/mpls/af_mpls.c
++++ b/net/mpls/af_mpls.c
+@@ -98,18 +98,19 @@ bool mpls_pkt_too_big(const struct sk_buff *skb, unsigned 
int mtu)
+ }
+ EXPORT_SYMBOL_GPL(mpls_pkt_too_big);
+ 
+-static u32 mpls_multipath_hash(struct mpls_route *rt,
+-                             struct sk_buff *skb, bool bos)
++static u32 mpls_multipath_hash(struct mpls_route *rt, struct sk_buff *skb)
+ {
+       struct mpls_entry_decoded dec;
++      unsigned int mpls_hdr_len = 0;
+       struct mpls_shim_hdr *hdr;
+       bool eli_seen = false;
+       int label_index;
+       u32 hash = 0;
+ 
+-      for (label_index = 0; label_index < MAX_MP_SELECT_LABELS && !bos;
++      for (label_index = 0; label_index < MAX_MP_SELECT_LABELS;
+            label_index++) {
+-              if (!pskb_may_pull(skb, sizeof(*hdr) * label_index))
++              mpls_hdr_len += sizeof(*hdr);
++              if (!pskb_may_pull(skb, mpls_hdr_len))
+                       break;
+ 
+               /* Read and decode the current label */
+@@ -134,37 +135,38 @@ static u32 mpls_multipath_hash(struct mpls_route *rt,
+                       eli_seen = true;
+               }
+ 
+-              bos = dec.bos;
+-              if (bos && pskb_may_pull(skb, sizeof(*hdr) * label_index +
+-                                       sizeof(struct iphdr))) {
++              if (!dec.bos)
++                      continue;
++
++              /* found bottom label; does skb have room for a header? */
++              if (pskb_may_pull(skb, mpls_hdr_len + sizeof(struct iphdr))) {
+                       const struct iphdr *v4hdr;
+ 
+-                      v4hdr = (const struct iphdr *)(mpls_hdr(skb) +
+-                                                     label_index);
++                      v4hdr = (const struct iphdr *)(hdr + 1);
+                       if (v4hdr->version == 4) {
+                               hash = jhash_3words(ntohl(v4hdr->saddr),
+                                                   ntohl(v4hdr->daddr),
+                                                   v4hdr->protocol, hash);
+                       } else if (v4hdr->version == 6 &&
+-                              pskb_may_pull(skb, sizeof(*hdr) * label_index +
+-                                            sizeof(struct ipv6hdr))) {
++                                 pskb_may_pull(skb, mpls_hdr_len +
++                                               sizeof(struct ipv6hdr))) {
+                               const struct ipv6hdr *v6hdr;
+ 
+-                              v6hdr = (const struct ipv6hdr *)(mpls_hdr(skb) +
+-                                                              label_index);
+-
++                              v6hdr = (const struct ipv6hdr *)(hdr + 1);
+                               hash = __ipv6_addr_jhash(&v6hdr->saddr, hash);
+                               hash = __ipv6_addr_jhash(&v6hdr->daddr, hash);
+                               hash = jhash_1word(v6hdr->nexthdr, hash);
+                       }
+               }
++
++              break;
+       }
+ 
+       return hash;
+ }
+ 
+ static struct mpls_nh *mpls_select_multipath(struct mpls_route *rt,
+-                                           struct sk_buff *skb, bool bos)
++                                           struct sk_buff *skb)
+ {
+       int alive = ACCESS_ONCE(rt->rt_nhn_alive);
+       u32 hash = 0;
+@@ -180,7 +182,7 @@ static struct mpls_nh *mpls_select_multipath(struct 
mpls_route *rt,
+       if (alive <= 0)
+               return NULL;
+ 
+-      hash = mpls_multipath_hash(rt, skb, bos);
++      hash = mpls_multipath_hash(rt, skb);
+       nh_index = hash % alive;
+       if (alive == rt->rt_nhn)
+               goto out;
+@@ -278,17 +280,11 @@ static int mpls_forward(struct sk_buff *skb, struct 
net_device *dev,
+       hdr = mpls_hdr(skb);
+       dec = mpls_entry_decode(hdr);
+ 
+-      /* Pop the label */
+-      skb_pull(skb, sizeof(*hdr));
+-      skb_reset_network_header(skb);
+-
+-      skb_orphan(skb);
+-
+       rt = mpls_route_input_rcu(net, dec.label);
+       if (!rt)
+               goto drop;
+ 
+-      nh = mpls_select_multipath(rt, skb, dec.bos);
++      nh = mpls_select_multipath(rt, skb);
+       if (!nh)
+               goto drop;
+ 
+@@ -297,6 +293,12 @@ static int mpls_forward(struct sk_buff *skb, struct 
net_device *dev,
+       if (!mpls_output_possible(out_dev))
+               goto drop;
+ 
++      /* Pop the label */
++      skb_pull(skb, sizeof(*hdr));
++      skb_reset_network_header(skb);
++
++      skb_orphan(skb);
++
+       if (skb_warn_if_lro(skb))
+               goto drop;
+ 
+diff --git a/net/mpls/mpls_iptunnel.c b/net/mpls/mpls_iptunnel.c
+index cf52cf30ac4b..bc9aaf58c7cc 100644
+--- a/net/mpls/mpls_iptunnel.c
++++ b/net/mpls/mpls_iptunnel.c
+@@ -218,6 +218,7 @@ static const struct lwtunnel_encap_ops mpls_iptun_ops = {
+       .fill_encap = mpls_fill_encap_info,
+       .get_encap_size = mpls_encap_nlsize,
+       .cmp_encap = mpls_encap_cmp,
++      .owner = THIS_MODULE,
+ };
+ 
+ static int __init mpls_iptunnel_init(void)
+diff --git a/net/openvswitch/conntrack.c b/net/openvswitch/conntrack.c
+index fecefa2dc94e..eab210bb1ef0 100644
+--- a/net/openvswitch/conntrack.c
++++ b/net/openvswitch/conntrack.c
+@@ -514,7 +514,7 @@ static int ovs_ct_nat_execute(struct sk_buff *skb, struct 
nf_conn *ct,
+       int hooknum, nh_off, err = NF_ACCEPT;
+ 
+       nh_off = skb_network_offset(skb);
+-      skb_pull(skb, nh_off);
++      skb_pull_rcsum(skb, nh_off);
+ 
+       /* See HOOK2MANIP(). */
+       if (maniptype == NF_NAT_MANIP_SRC)
+@@ -579,6 +579,7 @@ static int ovs_ct_nat_execute(struct sk_buff *skb, struct 
nf_conn *ct,
+       err = nf_nat_packet(ct, ctinfo, hooknum, skb);
+ push:
+       skb_push(skb, nh_off);
++      skb_postpush_rcsum(skb, skb->data, nh_off);
+ 
+       return err;
+ }
+@@ -890,7 +891,7 @@ int ovs_ct_execute(struct net *net, struct sk_buff *skb,
+ 
+       /* The conntrack module expects to be working at L3. */
+       nh_ofs = skb_network_offset(skb);
+-      skb_pull(skb, nh_ofs);
++      skb_pull_rcsum(skb, nh_ofs);
+ 
+       if (key->ip.frag != OVS_FRAG_TYPE_NONE) {
+               err = handle_fragments(net, key, info->zone.id, skb);
+@@ -904,6 +905,7 @@ int ovs_ct_execute(struct net *net, struct sk_buff *skb,
+               err = ovs_ct_lookup(net, key, info, skb);
+ 
+       skb_push(skb, nh_ofs);
++      skb_postpush_rcsum(skb, skb->data, nh_ofs);
+       if (err)
+               kfree_skb(skb);
+       return err;
+diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c
+index dd2332390c45..94e4a5941d89 100644
+--- a/net/packet/af_packet.c
++++ b/net/packet/af_packet.c
+@@ -1972,7 +1972,7 @@ static int __packet_rcv_vnet(const struct sk_buff *skb,
+ {
+       *vnet_hdr = (const struct virtio_net_hdr) { 0 };
+ 
+-      if (virtio_net_hdr_from_skb(skb, vnet_hdr, vio_le()))
++      if (virtio_net_hdr_from_skb(skb, vnet_hdr, vio_le(), true))
+               BUG();
+ 
+       return 0;
+diff --git a/net/sched/act_api.c b/net/sched/act_api.c
+index f893d180da1c..c6c2a93cc2a2 100644
+--- a/net/sched/act_api.c
++++ b/net/sched/act_api.c
+@@ -903,8 +903,6 @@ tca_action_gd(struct net *net, struct nlattr *nla, struct 
nlmsghdr *n,
+                       goto err;
+               }
+               act->order = i;
+-              if (event == RTM_GETACTION)
+-                      act->tcfa_refcnt++;
+               list_add_tail(&act->list, &actions);
+       }
+ 
+@@ -917,7 +915,8 @@ tca_action_gd(struct net *net, struct nlattr *nla, struct 
nlmsghdr *n,
+               return ret;
+       }
+ err:
+-      tcf_action_destroy(&actions, 0);
++      if (event != RTM_GETACTION)
++              tcf_action_destroy(&actions, 0);
+       return ret;
+ }
+ 
+diff --git a/net/unix/af_unix.c b/net/unix/af_unix.c
+index 2358f2690ec5..2d03d5bcb5b9 100644
+--- a/net/unix/af_unix.c
++++ b/net/unix/af_unix.c
+@@ -995,6 +995,7 @@ static int unix_bind(struct socket *sock, struct sockaddr 
*uaddr, int addr_len)
+       unsigned int hash;
+       struct unix_address *addr;
+       struct hlist_head *list;
++      struct path path = { NULL, NULL };
+ 
+       err = -EINVAL;
+       if (sunaddr->sun_family != AF_UNIX)
+@@ -1010,9 +1011,20 @@ static int unix_bind(struct socket *sock, struct 
sockaddr *uaddr, int addr_len)
+               goto out;
+       addr_len = err;
+ 
++      if (sun_path[0]) {
++              umode_t mode = S_IFSOCK |
++                     (SOCK_INODE(sock)->i_mode & ~current_umask());
++              err = unix_mknod(sun_path, mode, &path);
++              if (err) {
++                      if (err == -EEXIST)
++                              err = -EADDRINUSE;
++                      goto out;
++              }
++      }
++
+       err = mutex_lock_interruptible(&u->bindlock);
+       if (err)
+-              goto out;
++              goto out_put;
+ 
+       err = -EINVAL;
+       if (u->addr)
+@@ -1029,16 +1041,6 @@ static int unix_bind(struct socket *sock, struct 
sockaddr *uaddr, int addr_len)
+       atomic_set(&addr->refcnt, 1);
+ 
+       if (sun_path[0]) {
+-              struct path path;
+-              umode_t mode = S_IFSOCK |
+-                     (SOCK_INODE(sock)->i_mode & ~current_umask());
+-              err = unix_mknod(sun_path, mode, &path);
+-              if (err) {
+-                      if (err == -EEXIST)
+-                              err = -EADDRINUSE;
+-                      unix_release_addr(addr);
+-                      goto out_up;
+-              }
+               addr->hash = UNIX_HASH_SIZE;
+               hash = d_real_inode(path.dentry)->i_ino & (UNIX_HASH_SIZE - 1);
+               spin_lock(&unix_table_lock);
+@@ -1065,6 +1067,9 @@ static int unix_bind(struct socket *sock, struct 
sockaddr *uaddr, int addr_len)
+       spin_unlock(&unix_table_lock);
+ out_up:
+       mutex_unlock(&u->bindlock);
++out_put:
++      if (err)
++              path_put(&path);
+ out:
+       return err;
+ }

Reply via email to