]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
Merge tag 'for-next-dma_ops' of git://git.kernel.org/pub/scm/linux/kernel/git/dledfor...
authorLinus Torvalds <torvalds@linux-foundation.org>
Sat, 25 Feb 2017 21:45:43 +0000 (13:45 -0800)
committerLinus Torvalds <torvalds@linux-foundation.org>
Sat, 25 Feb 2017 21:45:43 +0000 (13:45 -0800)
Pull rdma DMA mapping updates from Doug Ledford:
 "Drop IB DMA mapping code and use core DMA code instead.

  Bart Van Assche noted that the ib DMA mapping code was significantly
  similar enough to the core DMA mapping code that with a few changes it
  was possible to remove the IB DMA mapping code entirely and switch the
  RDMA stack to use the core DMA mapping code.

  This resulted in a nice set of cleanups, but touched the entire tree
  and has been kept separate for that reason."

* tag 'for-next-dma_ops' of git://git.kernel.org/pub/scm/linux/kernel/git/dledford/rdma: (37 commits)
  IB/rxe, IB/rdmavt: Use dma_virt_ops instead of duplicating it
  IB/core: Remove ib_device.dma_device
  nvme-rdma: Switch from dma_device to dev.parent
  RDS: net: Switch from dma_device to dev.parent
  IB/srpt: Modify a debug statement
  IB/srp: Switch from dma_device to dev.parent
  IB/iser: Switch from dma_device to dev.parent
  IB/IPoIB: Switch from dma_device to dev.parent
  IB/rxe: Switch from dma_device to dev.parent
  IB/vmw_pvrdma: Switch from dma_device to dev.parent
  IB/usnic: Switch from dma_device to dev.parent
  IB/qib: Switch from dma_device to dev.parent
  IB/qedr: Switch from dma_device to dev.parent
  IB/ocrdma: Switch from dma_device to dev.parent
  IB/nes: Remove a superfluous assignment statement
  IB/mthca: Switch from dma_device to dev.parent
  IB/mlx5: Switch from dma_device to dev.parent
  IB/mlx4: Switch from dma_device to dev.parent
  IB/i40iw: Remove a superfluous assignment statement
  IB/hns: Switch from dma_device to dev.parent
  ...

45 files changed:
1  2 
arch/arm/mm/dma-mapping.c
arch/arm64/mm/dma-mapping.c
arch/m68k/kernel/dma.c
arch/mips/cavium-octeon/dma-octeon.c
arch/mips/loongson64/common/dma-swiotlb.c
arch/mips/mm/dma-default.c
arch/powerpc/kernel/pci-common.c
arch/s390/Kconfig
arch/s390/pci/pci.c
arch/x86/kernel/pci-calgary_64.c
arch/x86/kernel/pci-dma.c
arch/xtensa/kernel/pci-dma.c
drivers/infiniband/core/device.c
drivers/infiniband/hw/bnxt_re/main.c
drivers/infiniband/hw/cxgb3/iwch_provider.c
drivers/infiniband/hw/cxgb4/provider.c
drivers/infiniband/hw/hfi1/verbs.c
drivers/infiniband/hw/hns/hns_roce_main.c
drivers/infiniband/hw/i40iw/i40iw_verbs.c
drivers/infiniband/hw/mlx4/main.c
drivers/infiniband/hw/mlx5/main.c
drivers/infiniband/hw/mlx5/mr.c
drivers/infiniband/hw/mthca/mthca_provider.c
drivers/infiniband/hw/nes/nes_verbs.c
drivers/infiniband/hw/ocrdma/ocrdma_main.c
drivers/infiniband/hw/qedr/main.c
drivers/infiniband/hw/qib/qib_verbs.c
drivers/infiniband/hw/usnic/usnic_ib_main.c
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
drivers/infiniband/sw/rdmavt/Makefile
drivers/infiniband/sw/rdmavt/mr.c
drivers/infiniband/sw/rdmavt/vt.c
drivers/infiniband/sw/rxe/rxe_verbs.c
drivers/infiniband/ulp/ipoib/ipoib_main.c
drivers/infiniband/ulp/iser/iscsi_iser.c
drivers/infiniband/ulp/srp/ib_srp.c
drivers/infiniband/ulp/srpt/ib_srpt.c
drivers/iommu/amd_iommu.c
drivers/nvme/host/rdma.c
include/linux/device.h
include/linux/dma-mapping.h
include/rdma/ib_verbs.h
lib/Kconfig
lib/Makefile
net/rds/ib.h

Simple merge
index aff1d0afeb1eaded6f85bf8398b2e793cc104524,dbab4c6c084b39ecdfd767d3342d6848c21f58d1..81cdb2e844ed9fe80e4192315647ccc9245b8003
@@@ -356,14 -352,7 +356,14 @@@ static int __swiotlb_dma_supported(stru
        return 1;
  }
  
- static struct dma_map_ops swiotlb_dma_ops = {
 +static int __swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t addr)
 +{
 +      if (swiotlb)
 +              return swiotlb_dma_mapping_error(hwdev, addr);
 +      return 0;
 +}
 +
+ static const struct dma_map_ops swiotlb_dma_ops = {
        .alloc = __dma_alloc,
        .free = __dma_free,
        .mmap = __swiotlb_mmap,
@@@ -841,21 -831,14 +841,21 @@@ static bool do_iommu_attach(struct devi
         * then the IOMMU core will have already configured a group for this
         * device, and allocated the default domain for that group.
         */
 -      if (!domain || iommu_dma_init_domain(domain, dma_base, size, dev)) {
 -              pr_warn("Failed to set up IOMMU for device %s; retaining platform DMA ops\n",
 -                      dev_name(dev));
 -              return false;
 +      if (!domain)
 +              goto out_err;
 +
 +      if (domain->type == IOMMU_DOMAIN_DMA) {
 +              if (iommu_dma_init_domain(domain, dma_base, size, dev))
 +                      goto out_err;
 +
-               dev->archdata.dma_ops = &iommu_dma_ops;
++              dev->dma_ops = &iommu_dma_ops;
        }
  
 -      dev->dma_ops = &iommu_dma_ops;
        return true;
 +out_err:
 +      pr_warn("Failed to set up IOMMU for device %s; retaining platform DMA ops\n",
 +               dev_name(dev));
 +      return false;
  }
  
  static void queue_iommu_attach(struct device *dev, const struct iommu_ops *ops,
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
index bd452a92b386dfe4bcde2dd64c3fd97b0a0f3081,0000000000000000000000000000000000000000..5d355401179b8ae5107e411421c673db86164696
mode 100644,000000..100644
--- /dev/null
@@@ -1,1315 -1,0 +1,1315 @@@
-       ibdev->dma_device = &rdev->en_dev->pdev->dev;
 +/*
 + * Broadcom NetXtreme-E RoCE driver.
 + *
 + * Copyright (c) 2016 - 2017, Broadcom. All rights reserved.  The term
 + * Broadcom refers to Broadcom Limited and/or its subsidiaries.
 + *
 + * This software is available to you under a choice of one of two
 + * licenses.  You may choose to be licensed under the terms of the GNU
 + * General Public License (GPL) Version 2, available from the file
 + * COPYING in the main directory of this source tree, or the
 + * BSD license below:
 + *
 + * Redistribution and use in source and binary forms, with or without
 + * modification, are permitted provided that the following conditions
 + * are met:
 + *
 + * 1. Redistributions of source code must retain the above copyright
 + *    notice, this list of conditions and the following disclaimer.
 + * 2. Redistributions in binary form must reproduce the above copyright
 + *    notice, this list of conditions and the following disclaimer in
 + *    the documentation and/or other materials provided with the
 + *    distribution.
 + *
 + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS''
 + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
 + * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
 + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS
 + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
 + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
 + * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
 + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
 + * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
 + * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 + *
 + * Description: Main component of the bnxt_re driver
 + */
 +
 +#include <linux/module.h>
 +#include <linux/netdevice.h>
 +#include <linux/ethtool.h>
 +#include <linux/mutex.h>
 +#include <linux/list.h>
 +#include <linux/rculist.h>
 +#include <linux/spinlock.h>
 +#include <linux/pci.h>
 +#include <net/dcbnl.h>
 +#include <net/ipv6.h>
 +#include <net/addrconf.h>
 +#include <linux/if_ether.h>
 +
 +#include <rdma/ib_verbs.h>
 +#include <rdma/ib_user_verbs.h>
 +#include <rdma/ib_umem.h>
 +#include <rdma/ib_addr.h>
 +
 +#include "bnxt_ulp.h"
 +#include "roce_hsi.h"
 +#include "qplib_res.h"
 +#include "qplib_sp.h"
 +#include "qplib_fp.h"
 +#include "qplib_rcfw.h"
 +#include "bnxt_re.h"
 +#include "ib_verbs.h"
 +#include <rdma/bnxt_re-abi.h>
 +#include "bnxt.h"
 +static char version[] =
 +              BNXT_RE_DESC " v" ROCE_DRV_MODULE_VERSION "\n";
 +
 +MODULE_AUTHOR("Eddie Wai <eddie.wai@broadcom.com>");
 +MODULE_DESCRIPTION(BNXT_RE_DESC " Driver");
 +MODULE_LICENSE("Dual BSD/GPL");
 +MODULE_VERSION(ROCE_DRV_MODULE_VERSION);
 +
 +/* globals */
 +static struct list_head bnxt_re_dev_list = LIST_HEAD_INIT(bnxt_re_dev_list);
 +/* Mutex to protect the list of bnxt_re devices added */
 +static DEFINE_MUTEX(bnxt_re_dev_lock);
 +static struct workqueue_struct *bnxt_re_wq;
 +
 +/* for handling bnxt_en callbacks later */
 +static void bnxt_re_stop(void *p)
 +{
 +}
 +
 +static void bnxt_re_start(void *p)
 +{
 +}
 +
 +static void bnxt_re_sriov_config(void *p, int num_vfs)
 +{
 +}
 +
 +static struct bnxt_ulp_ops bnxt_re_ulp_ops = {
 +      .ulp_async_notifier = NULL,
 +      .ulp_stop = bnxt_re_stop,
 +      .ulp_start = bnxt_re_start,
 +      .ulp_sriov_config = bnxt_re_sriov_config
 +};
 +
 +/* RoCE -> Net driver */
 +
 +/* Driver registration routines used to let the networking driver (bnxt_en)
 + * to know that the RoCE driver is now installed
 + */
 +static int bnxt_re_unregister_netdev(struct bnxt_re_dev *rdev, bool lock_wait)
 +{
 +      struct bnxt_en_dev *en_dev;
 +      int rc;
 +
 +      if (!rdev)
 +              return -EINVAL;
 +
 +      en_dev = rdev->en_dev;
 +      /* Acquire rtnl lock if it is not invokded from netdev event */
 +      if (lock_wait)
 +              rtnl_lock();
 +
 +      rc = en_dev->en_ops->bnxt_unregister_device(rdev->en_dev,
 +                                                  BNXT_ROCE_ULP);
 +      if (lock_wait)
 +              rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_register_netdev(struct bnxt_re_dev *rdev)
 +{
 +      struct bnxt_en_dev *en_dev;
 +      int rc = 0;
 +
 +      if (!rdev)
 +              return -EINVAL;
 +
 +      en_dev = rdev->en_dev;
 +
 +      rtnl_lock();
 +      rc = en_dev->en_ops->bnxt_register_device(en_dev, BNXT_ROCE_ULP,
 +                                                &bnxt_re_ulp_ops, rdev);
 +      rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_free_msix(struct bnxt_re_dev *rdev, bool lock_wait)
 +{
 +      struct bnxt_en_dev *en_dev;
 +      int rc;
 +
 +      if (!rdev)
 +              return -EINVAL;
 +
 +      en_dev = rdev->en_dev;
 +
 +      if (lock_wait)
 +              rtnl_lock();
 +
 +      rc = en_dev->en_ops->bnxt_free_msix(rdev->en_dev, BNXT_ROCE_ULP);
 +
 +      if (lock_wait)
 +              rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_request_msix(struct bnxt_re_dev *rdev)
 +{
 +      int rc = 0, num_msix_want = BNXT_RE_MIN_MSIX, num_msix_got;
 +      struct bnxt_en_dev *en_dev;
 +
 +      if (!rdev)
 +              return -EINVAL;
 +
 +      en_dev = rdev->en_dev;
 +
 +      rtnl_lock();
 +      num_msix_got = en_dev->en_ops->bnxt_request_msix(en_dev, BNXT_ROCE_ULP,
 +                                                       rdev->msix_entries,
 +                                                       num_msix_want);
 +      if (num_msix_got < BNXT_RE_MIN_MSIX) {
 +              rc = -EINVAL;
 +              goto done;
 +      }
 +      if (num_msix_got != num_msix_want) {
 +              dev_warn(rdev_to_dev(rdev),
 +                       "Requested %d MSI-X vectors, got %d\n",
 +                       num_msix_want, num_msix_got);
 +      }
 +      rdev->num_msix = num_msix_got;
 +done:
 +      rtnl_unlock();
 +      return rc;
 +}
 +
 +static void bnxt_re_init_hwrm_hdr(struct bnxt_re_dev *rdev, struct input *hdr,
 +                                u16 opcd, u16 crid, u16 trid)
 +{
 +      hdr->req_type = cpu_to_le16(opcd);
 +      hdr->cmpl_ring = cpu_to_le16(crid);
 +      hdr->target_id = cpu_to_le16(trid);
 +}
 +
 +static void bnxt_re_fill_fw_msg(struct bnxt_fw_msg *fw_msg, void *msg,
 +                              int msg_len, void *resp, int resp_max_len,
 +                              int timeout)
 +{
 +      fw_msg->msg = msg;
 +      fw_msg->msg_len = msg_len;
 +      fw_msg->resp = resp;
 +      fw_msg->resp_max_len = resp_max_len;
 +      fw_msg->timeout = timeout;
 +}
 +
 +static int bnxt_re_net_ring_free(struct bnxt_re_dev *rdev, u16 fw_ring_id,
 +                               bool lock_wait)
 +{
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct hwrm_ring_free_input req = {0};
 +      struct hwrm_ring_free_output resp;
 +      struct bnxt_fw_msg fw_msg;
 +      bool do_unlock = false;
 +      int rc = -EINVAL;
 +
 +      if (!en_dev)
 +              return rc;
 +
 +      memset(&fw_msg, 0, sizeof(fw_msg));
 +      if (lock_wait) {
 +              rtnl_lock();
 +              do_unlock = true;
 +      }
 +
 +      bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_RING_FREE, -1, -1);
 +      req.ring_type = RING_ALLOC_REQ_RING_TYPE_L2_CMPL;
 +      req.ring_id = cpu_to_le16(fw_ring_id);
 +      bnxt_re_fill_fw_msg(&fw_msg, (void *)&req, sizeof(req), (void *)&resp,
 +                          sizeof(resp), DFLT_HWRM_CMD_TIMEOUT);
 +      rc = en_dev->en_ops->bnxt_send_fw_msg(en_dev, BNXT_ROCE_ULP, &fw_msg);
 +      if (rc)
 +              dev_err(rdev_to_dev(rdev),
 +                      "Failed to free HW ring:%d :%#x", req.ring_id, rc);
 +      if (do_unlock)
 +              rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_net_ring_alloc(struct bnxt_re_dev *rdev, dma_addr_t *dma_arr,
 +                                int pages, int type, u32 ring_mask,
 +                                u32 map_index, u16 *fw_ring_id)
 +{
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct hwrm_ring_alloc_input req = {0};
 +      struct hwrm_ring_alloc_output resp;
 +      struct bnxt_fw_msg fw_msg;
 +      int rc = -EINVAL;
 +
 +      if (!en_dev)
 +              return rc;
 +
 +      memset(&fw_msg, 0, sizeof(fw_msg));
 +      rtnl_lock();
 +      bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_RING_ALLOC, -1, -1);
 +      req.enables = 0;
 +      req.page_tbl_addr =  cpu_to_le64(dma_arr[0]);
 +      if (pages > 1) {
 +              /* Page size is in log2 units */
 +              req.page_size = BNXT_PAGE_SHIFT;
 +              req.page_tbl_depth = 1;
 +      }
 +      req.fbo = 0;
 +      /* Association of ring index with doorbell index and MSIX number */
 +      req.logical_id = cpu_to_le16(map_index);
 +      req.length = cpu_to_le32(ring_mask + 1);
 +      req.ring_type = RING_ALLOC_REQ_RING_TYPE_L2_CMPL;
 +      req.int_mode = RING_ALLOC_REQ_INT_MODE_MSIX;
 +      bnxt_re_fill_fw_msg(&fw_msg, (void *)&req, sizeof(req), (void *)&resp,
 +                          sizeof(resp), DFLT_HWRM_CMD_TIMEOUT);
 +      rc = en_dev->en_ops->bnxt_send_fw_msg(en_dev, BNXT_ROCE_ULP, &fw_msg);
 +      if (!rc)
 +              *fw_ring_id = le16_to_cpu(resp.ring_id);
 +
 +      rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_net_stats_ctx_free(struct bnxt_re_dev *rdev,
 +                                    u32 fw_stats_ctx_id, bool lock_wait)
 +{
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct hwrm_stat_ctx_free_input req = {0};
 +      struct bnxt_fw_msg fw_msg;
 +      bool do_unlock = false;
 +      int rc = -EINVAL;
 +
 +      if (!en_dev)
 +              return rc;
 +
 +      memset(&fw_msg, 0, sizeof(fw_msg));
 +      if (lock_wait) {
 +              rtnl_lock();
 +              do_unlock = true;
 +      }
 +
 +      bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_STAT_CTX_FREE, -1, -1);
 +      req.stat_ctx_id = cpu_to_le32(fw_stats_ctx_id);
 +      bnxt_re_fill_fw_msg(&fw_msg, (void *)&req, sizeof(req), (void *)&req,
 +                          sizeof(req), DFLT_HWRM_CMD_TIMEOUT);
 +      rc = en_dev->en_ops->bnxt_send_fw_msg(en_dev, BNXT_ROCE_ULP, &fw_msg);
 +      if (rc)
 +              dev_err(rdev_to_dev(rdev),
 +                      "Failed to free HW stats context %#x", rc);
 +
 +      if (do_unlock)
 +              rtnl_unlock();
 +      return rc;
 +}
 +
 +static int bnxt_re_net_stats_ctx_alloc(struct bnxt_re_dev *rdev,
 +                                     dma_addr_t dma_map,
 +                                     u32 *fw_stats_ctx_id)
 +{
 +      struct hwrm_stat_ctx_alloc_output resp = {0};
 +      struct hwrm_stat_ctx_alloc_input req = {0};
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct bnxt_fw_msg fw_msg;
 +      int rc = -EINVAL;
 +
 +      *fw_stats_ctx_id = INVALID_STATS_CTX_ID;
 +
 +      if (!en_dev)
 +              return rc;
 +
 +      memset(&fw_msg, 0, sizeof(fw_msg));
 +      rtnl_lock();
 +
 +      bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_STAT_CTX_ALLOC, -1, -1);
 +      req.update_period_ms = cpu_to_le32(1000);
 +      req.stats_dma_addr = cpu_to_le64(dma_map);
 +      bnxt_re_fill_fw_msg(&fw_msg, (void *)&req, sizeof(req), (void *)&resp,
 +                          sizeof(resp), DFLT_HWRM_CMD_TIMEOUT);
 +      rc = en_dev->en_ops->bnxt_send_fw_msg(en_dev, BNXT_ROCE_ULP, &fw_msg);
 +      if (!rc)
 +              *fw_stats_ctx_id = le32_to_cpu(resp.stat_ctx_id);
 +
 +      rtnl_unlock();
 +      return rc;
 +}
 +
 +/* Device */
 +
 +static bool is_bnxt_re_dev(struct net_device *netdev)
 +{
 +      struct ethtool_drvinfo drvinfo;
 +
 +      if (netdev->ethtool_ops && netdev->ethtool_ops->get_drvinfo) {
 +              memset(&drvinfo, 0, sizeof(drvinfo));
 +              netdev->ethtool_ops->get_drvinfo(netdev, &drvinfo);
 +
 +              if (strcmp(drvinfo.driver, "bnxt_en"))
 +                      return false;
 +              return true;
 +      }
 +      return false;
 +}
 +
 +static struct bnxt_re_dev *bnxt_re_from_netdev(struct net_device *netdev)
 +{
 +      struct bnxt_re_dev *rdev;
 +
 +      rcu_read_lock();
 +      list_for_each_entry_rcu(rdev, &bnxt_re_dev_list, list) {
 +              if (rdev->netdev == netdev) {
 +                      rcu_read_unlock();
 +                      return rdev;
 +              }
 +      }
 +      rcu_read_unlock();
 +      return NULL;
 +}
 +
 +static void bnxt_re_dev_unprobe(struct net_device *netdev,
 +                              struct bnxt_en_dev *en_dev)
 +{
 +      dev_put(netdev);
 +      module_put(en_dev->pdev->driver->driver.owner);
 +}
 +
 +static struct bnxt_en_dev *bnxt_re_dev_probe(struct net_device *netdev)
 +{
 +      struct bnxt *bp = netdev_priv(netdev);
 +      struct bnxt_en_dev *en_dev;
 +      struct pci_dev *pdev;
 +
 +      /* Call bnxt_en's RoCE probe via indirect API */
 +      if (!bp->ulp_probe)
 +              return ERR_PTR(-EINVAL);
 +
 +      en_dev = bp->ulp_probe(netdev);
 +      if (IS_ERR(en_dev))
 +              return en_dev;
 +
 +      pdev = en_dev->pdev;
 +      if (!pdev)
 +              return ERR_PTR(-EINVAL);
 +
 +      if (!(en_dev->flags & BNXT_EN_FLAG_ROCE_CAP)) {
 +              dev_dbg(&pdev->dev,
 +                      "%s: probe error: RoCE is not supported on this device",
 +                      ROCE_DRV_MODULE_NAME);
 +              return ERR_PTR(-ENODEV);
 +      }
 +
 +      /* Bump net device reference count */
 +      if (!try_module_get(pdev->driver->driver.owner))
 +              return ERR_PTR(-ENODEV);
 +
 +      dev_hold(netdev);
 +
 +      return en_dev;
 +}
 +
 +static void bnxt_re_unregister_ib(struct bnxt_re_dev *rdev)
 +{
 +      ib_unregister_device(&rdev->ibdev);
 +}
 +
 +static int bnxt_re_register_ib(struct bnxt_re_dev *rdev)
 +{
 +      struct ib_device *ibdev = &rdev->ibdev;
 +
 +      /* ib device init */
 +      ibdev->owner = THIS_MODULE;
 +      ibdev->node_type = RDMA_NODE_IB_CA;
 +      strlcpy(ibdev->name, "bnxt_re%d", IB_DEVICE_NAME_MAX);
 +      strlcpy(ibdev->node_desc, BNXT_RE_DESC " HCA",
 +              strlen(BNXT_RE_DESC) + 5);
 +      ibdev->phys_port_cnt = 1;
 +
 +      bnxt_qplib_get_guid(rdev->netdev->dev_addr, (u8 *)&ibdev->node_guid);
 +
 +      ibdev->num_comp_vectors = 1;
++      ibdev->dev.parent = &rdev->en_dev->pdev->dev;
 +      ibdev->local_dma_lkey = BNXT_QPLIB_RSVD_LKEY;
 +
 +      /* User space */
 +      ibdev->uverbs_abi_ver = BNXT_RE_ABI_VERSION;
 +      ibdev->uverbs_cmd_mask =
 +                      (1ull << IB_USER_VERBS_CMD_GET_CONTEXT)         |
 +                      (1ull << IB_USER_VERBS_CMD_QUERY_DEVICE)        |
 +                      (1ull << IB_USER_VERBS_CMD_QUERY_PORT)          |
 +                      (1ull << IB_USER_VERBS_CMD_ALLOC_PD)            |
 +                      (1ull << IB_USER_VERBS_CMD_DEALLOC_PD)          |
 +                      (1ull << IB_USER_VERBS_CMD_REG_MR)              |
 +                      (1ull << IB_USER_VERBS_CMD_REREG_MR)            |
 +                      (1ull << IB_USER_VERBS_CMD_DEREG_MR)            |
 +                      (1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
 +                      (1ull << IB_USER_VERBS_CMD_CREATE_CQ)           |
 +                      (1ull << IB_USER_VERBS_CMD_RESIZE_CQ)           |
 +                      (1ull << IB_USER_VERBS_CMD_DESTROY_CQ)          |
 +                      (1ull << IB_USER_VERBS_CMD_CREATE_QP)           |
 +                      (1ull << IB_USER_VERBS_CMD_MODIFY_QP)           |
 +                      (1ull << IB_USER_VERBS_CMD_QUERY_QP)            |
 +                      (1ull << IB_USER_VERBS_CMD_DESTROY_QP)          |
 +                      (1ull << IB_USER_VERBS_CMD_CREATE_SRQ)          |
 +                      (1ull << IB_USER_VERBS_CMD_MODIFY_SRQ)          |
 +                      (1ull << IB_USER_VERBS_CMD_QUERY_SRQ)           |
 +                      (1ull << IB_USER_VERBS_CMD_DESTROY_SRQ)         |
 +                      (1ull << IB_USER_VERBS_CMD_CREATE_AH)           |
 +                      (1ull << IB_USER_VERBS_CMD_MODIFY_AH)           |
 +                      (1ull << IB_USER_VERBS_CMD_QUERY_AH)            |
 +                      (1ull << IB_USER_VERBS_CMD_DESTROY_AH);
 +      /* POLL_CQ and REQ_NOTIFY_CQ is directly handled in libbnxt_re */
 +
 +      /* Kernel verbs */
 +      ibdev->query_device             = bnxt_re_query_device;
 +      ibdev->modify_device            = bnxt_re_modify_device;
 +
 +      ibdev->query_port               = bnxt_re_query_port;
 +      ibdev->modify_port              = bnxt_re_modify_port;
 +      ibdev->get_port_immutable       = bnxt_re_get_port_immutable;
 +      ibdev->query_pkey               = bnxt_re_query_pkey;
 +      ibdev->query_gid                = bnxt_re_query_gid;
 +      ibdev->get_netdev               = bnxt_re_get_netdev;
 +      ibdev->add_gid                  = bnxt_re_add_gid;
 +      ibdev->del_gid                  = bnxt_re_del_gid;
 +      ibdev->get_link_layer           = bnxt_re_get_link_layer;
 +
 +      ibdev->alloc_pd                 = bnxt_re_alloc_pd;
 +      ibdev->dealloc_pd               = bnxt_re_dealloc_pd;
 +
 +      ibdev->create_ah                = bnxt_re_create_ah;
 +      ibdev->modify_ah                = bnxt_re_modify_ah;
 +      ibdev->query_ah                 = bnxt_re_query_ah;
 +      ibdev->destroy_ah               = bnxt_re_destroy_ah;
 +
 +      ibdev->create_qp                = bnxt_re_create_qp;
 +      ibdev->modify_qp                = bnxt_re_modify_qp;
 +      ibdev->query_qp                 = bnxt_re_query_qp;
 +      ibdev->destroy_qp               = bnxt_re_destroy_qp;
 +
 +      ibdev->post_send                = bnxt_re_post_send;
 +      ibdev->post_recv                = bnxt_re_post_recv;
 +
 +      ibdev->create_cq                = bnxt_re_create_cq;
 +      ibdev->destroy_cq               = bnxt_re_destroy_cq;
 +      ibdev->poll_cq                  = bnxt_re_poll_cq;
 +      ibdev->req_notify_cq            = bnxt_re_req_notify_cq;
 +
 +      ibdev->get_dma_mr               = bnxt_re_get_dma_mr;
 +      ibdev->dereg_mr                 = bnxt_re_dereg_mr;
 +      ibdev->alloc_mr                 = bnxt_re_alloc_mr;
 +      ibdev->map_mr_sg                = bnxt_re_map_mr_sg;
 +      ibdev->alloc_fmr                = bnxt_re_alloc_fmr;
 +      ibdev->map_phys_fmr             = bnxt_re_map_phys_fmr;
 +      ibdev->unmap_fmr                = bnxt_re_unmap_fmr;
 +      ibdev->dealloc_fmr              = bnxt_re_dealloc_fmr;
 +
 +      ibdev->reg_user_mr              = bnxt_re_reg_user_mr;
 +      ibdev->alloc_ucontext           = bnxt_re_alloc_ucontext;
 +      ibdev->dealloc_ucontext         = bnxt_re_dealloc_ucontext;
 +      ibdev->mmap                     = bnxt_re_mmap;
 +
 +      return ib_register_device(ibdev, NULL);
 +}
 +
 +static ssize_t show_rev(struct device *device, struct device_attribute *attr,
 +                      char *buf)
 +{
 +      struct bnxt_re_dev *rdev = to_bnxt_re_dev(device, ibdev.dev);
 +
 +      return scnprintf(buf, PAGE_SIZE, "0x%x\n", rdev->en_dev->pdev->vendor);
 +}
 +
 +static ssize_t show_fw_ver(struct device *device, struct device_attribute *attr,
 +                         char *buf)
 +{
 +      struct bnxt_re_dev *rdev = to_bnxt_re_dev(device, ibdev.dev);
 +
 +      return scnprintf(buf, PAGE_SIZE, "%s\n", rdev->dev_attr.fw_ver);
 +}
 +
 +static ssize_t show_hca(struct device *device, struct device_attribute *attr,
 +                      char *buf)
 +{
 +      struct bnxt_re_dev *rdev = to_bnxt_re_dev(device, ibdev.dev);
 +
 +      return scnprintf(buf, PAGE_SIZE, "%s\n", rdev->ibdev.node_desc);
 +}
 +
 +static DEVICE_ATTR(hw_rev, 0444, show_rev, NULL);
 +static DEVICE_ATTR(fw_rev, 0444, show_fw_ver, NULL);
 +static DEVICE_ATTR(hca_type, 0444, show_hca, NULL);
 +
 +static struct device_attribute *bnxt_re_attributes[] = {
 +      &dev_attr_hw_rev,
 +      &dev_attr_fw_rev,
 +      &dev_attr_hca_type
 +};
 +
 +static void bnxt_re_dev_remove(struct bnxt_re_dev *rdev)
 +{
 +      dev_put(rdev->netdev);
 +      rdev->netdev = NULL;
 +
 +      mutex_lock(&bnxt_re_dev_lock);
 +      list_del_rcu(&rdev->list);
 +      mutex_unlock(&bnxt_re_dev_lock);
 +
 +      synchronize_rcu();
 +      flush_workqueue(bnxt_re_wq);
 +
 +      ib_dealloc_device(&rdev->ibdev);
 +      /* rdev is gone */
 +}
 +
 +static struct bnxt_re_dev *bnxt_re_dev_add(struct net_device *netdev,
 +                                         struct bnxt_en_dev *en_dev)
 +{
 +      struct bnxt_re_dev *rdev;
 +
 +      /* Allocate bnxt_re_dev instance here */
 +      rdev = (struct bnxt_re_dev *)ib_alloc_device(sizeof(*rdev));
 +      if (!rdev) {
 +              dev_err(NULL, "%s: bnxt_re_dev allocation failure!",
 +                      ROCE_DRV_MODULE_NAME);
 +              return NULL;
 +      }
 +      /* Default values */
 +      rdev->netdev = netdev;
 +      dev_hold(rdev->netdev);
 +      rdev->en_dev = en_dev;
 +      rdev->id = rdev->en_dev->pdev->devfn;
 +      INIT_LIST_HEAD(&rdev->qp_list);
 +      mutex_init(&rdev->qp_lock);
 +      atomic_set(&rdev->qp_count, 0);
 +      atomic_set(&rdev->cq_count, 0);
 +      atomic_set(&rdev->srq_count, 0);
 +      atomic_set(&rdev->mr_count, 0);
 +      atomic_set(&rdev->mw_count, 0);
 +      rdev->cosq[0] = 0xFFFF;
 +      rdev->cosq[1] = 0xFFFF;
 +
 +      mutex_lock(&bnxt_re_dev_lock);
 +      list_add_tail_rcu(&rdev->list, &bnxt_re_dev_list);
 +      mutex_unlock(&bnxt_re_dev_lock);
 +      return rdev;
 +}
 +
 +static int bnxt_re_aeq_handler(struct bnxt_qplib_rcfw *rcfw,
 +                             struct creq_func_event *aeqe)
 +{
 +      switch (aeqe->event) {
 +      case CREQ_FUNC_EVENT_EVENT_TX_WQE_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_TX_DATA_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_RX_WQE_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_RX_DATA_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_CQ_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_TQM_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_CFCQ_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_CFCS_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_CFCC_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_CFCM_ERROR:
 +              break;
 +      case CREQ_FUNC_EVENT_EVENT_TIM_ERROR:
 +              break;
 +      default:
 +              return -EINVAL;
 +      }
 +      return 0;
 +}
 +
 +static int bnxt_re_cqn_handler(struct bnxt_qplib_nq *nq,
 +                             struct bnxt_qplib_cq *handle)
 +{
 +      struct bnxt_re_cq *cq = container_of(handle, struct bnxt_re_cq,
 +                                           qplib_cq);
 +
 +      if (!cq) {
 +              dev_err(NULL, "%s: CQ is NULL, CQN not handled",
 +                      ROCE_DRV_MODULE_NAME);
 +              return -EINVAL;
 +      }
 +      if (cq->ib_cq.comp_handler) {
 +              /* Lock comp_handler? */
 +              (*cq->ib_cq.comp_handler)(&cq->ib_cq, cq->ib_cq.cq_context);
 +      }
 +
 +      return 0;
 +}
 +
 +static void bnxt_re_cleanup_res(struct bnxt_re_dev *rdev)
 +{
 +      if (rdev->nq.hwq.max_elements)
 +              bnxt_qplib_disable_nq(&rdev->nq);
 +
 +      if (rdev->qplib_res.rcfw)
 +              bnxt_qplib_cleanup_res(&rdev->qplib_res);
 +}
 +
 +static int bnxt_re_init_res(struct bnxt_re_dev *rdev)
 +{
 +      int rc = 0;
 +
 +      bnxt_qplib_init_res(&rdev->qplib_res);
 +
 +      if (rdev->msix_entries[BNXT_RE_NQ_IDX].vector <= 0)
 +              return -EINVAL;
 +
 +      rc = bnxt_qplib_enable_nq(rdev->en_dev->pdev, &rdev->nq,
 +                                rdev->msix_entries[BNXT_RE_NQ_IDX].vector,
 +                                rdev->msix_entries[BNXT_RE_NQ_IDX].db_offset,
 +                                &bnxt_re_cqn_handler,
 +                                NULL);
 +
 +      if (rc)
 +              dev_err(rdev_to_dev(rdev), "Failed to enable NQ: %#x", rc);
 +
 +      return rc;
 +}
 +
 +static void bnxt_re_free_res(struct bnxt_re_dev *rdev, bool lock_wait)
 +{
 +      if (rdev->nq.hwq.max_elements) {
 +              bnxt_re_net_ring_free(rdev, rdev->nq.ring_id, lock_wait);
 +              bnxt_qplib_free_nq(&rdev->nq);
 +      }
 +      if (rdev->qplib_res.dpi_tbl.max) {
 +              bnxt_qplib_dealloc_dpi(&rdev->qplib_res,
 +                                     &rdev->qplib_res.dpi_tbl,
 +                                     &rdev->dpi_privileged);
 +      }
 +      if (rdev->qplib_res.rcfw) {
 +              bnxt_qplib_free_res(&rdev->qplib_res);
 +              rdev->qplib_res.rcfw = NULL;
 +      }
 +}
 +
 +static int bnxt_re_alloc_res(struct bnxt_re_dev *rdev)
 +{
 +      int rc = 0;
 +
 +      /* Configure and allocate resources for qplib */
 +      rdev->qplib_res.rcfw = &rdev->rcfw;
 +      rc = bnxt_qplib_get_dev_attr(&rdev->rcfw, &rdev->dev_attr);
 +      if (rc)
 +              goto fail;
 +
 +      rc = bnxt_qplib_alloc_res(&rdev->qplib_res, rdev->en_dev->pdev,
 +                                rdev->netdev, &rdev->dev_attr);
 +      if (rc)
 +              goto fail;
 +
 +      rc = bnxt_qplib_alloc_dpi(&rdev->qplib_res.dpi_tbl,
 +                                &rdev->dpi_privileged,
 +                                rdev);
 +      if (rc)
 +              goto fail;
 +
 +      rdev->nq.hwq.max_elements = BNXT_RE_MAX_CQ_COUNT +
 +                                  BNXT_RE_MAX_SRQC_COUNT + 2;
 +      rc = bnxt_qplib_alloc_nq(rdev->en_dev->pdev, &rdev->nq);
 +      if (rc) {
 +              dev_err(rdev_to_dev(rdev),
 +                      "Failed to allocate NQ memory: %#x", rc);
 +              goto fail;
 +      }
 +      rc = bnxt_re_net_ring_alloc
 +                      (rdev, rdev->nq.hwq.pbl[PBL_LVL_0].pg_map_arr,
 +                       rdev->nq.hwq.pbl[rdev->nq.hwq.level].pg_count,
 +                       HWRM_RING_ALLOC_CMPL, BNXT_QPLIB_NQE_MAX_CNT - 1,
 +                       rdev->msix_entries[BNXT_RE_NQ_IDX].ring_idx,
 +                       &rdev->nq.ring_id);
 +      if (rc) {
 +              dev_err(rdev_to_dev(rdev),
 +                      "Failed to allocate NQ ring: %#x", rc);
 +              goto free_nq;
 +      }
 +      return 0;
 +free_nq:
 +      bnxt_qplib_free_nq(&rdev->nq);
 +fail:
 +      rdev->qplib_res.rcfw = NULL;
 +      return rc;
 +}
 +
 +static void bnxt_re_dispatch_event(struct ib_device *ibdev, struct ib_qp *qp,
 +                                 u8 port_num, enum ib_event_type event)
 +{
 +      struct ib_event ib_event;
 +
 +      ib_event.device = ibdev;
 +      if (qp)
 +              ib_event.element.qp = qp;
 +      else
 +              ib_event.element.port_num = port_num;
 +      ib_event.event = event;
 +      ib_dispatch_event(&ib_event);
 +}
 +
 +#define HWRM_QUEUE_PRI2COS_QCFG_INPUT_FLAGS_IVLAN      0x02
 +static int bnxt_re_query_hwrm_pri2cos(struct bnxt_re_dev *rdev, u8 dir,
 +                                    u64 *cid_map)
 +{
 +      struct hwrm_queue_pri2cos_qcfg_input req = {0};
 +      struct bnxt *bp = netdev_priv(rdev->netdev);
 +      struct hwrm_queue_pri2cos_qcfg_output resp;
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct bnxt_fw_msg fw_msg;
 +      u32 flags = 0;
 +      u8 *qcfgmap, *tmp_map;
 +      int rc = 0, i;
 +
 +      if (!cid_map)
 +              return -EINVAL;
 +
 +      memset(&fw_msg, 0, sizeof(fw_msg));
 +      bnxt_re_init_hwrm_hdr(rdev, (void *)&req,
 +                            HWRM_QUEUE_PRI2COS_QCFG, -1, -1);
 +      flags |= (dir & 0x01);
 +      flags |= HWRM_QUEUE_PRI2COS_QCFG_INPUT_FLAGS_IVLAN;
 +      req.flags = cpu_to_le32(flags);
 +      req.port_id = bp->pf.port_id;
 +
 +      bnxt_re_fill_fw_msg(&fw_msg, (void *)&req, sizeof(req), (void *)&resp,
 +                          sizeof(resp), DFLT_HWRM_CMD_TIMEOUT);
 +      rc = en_dev->en_ops->bnxt_send_fw_msg(en_dev, BNXT_ROCE_ULP, &fw_msg);
 +      if (rc)
 +              return rc;
 +
 +      if (resp.queue_cfg_info) {
 +              dev_warn(rdev_to_dev(rdev),
 +                       "Asymmetric cos queue configuration detected");
 +              dev_warn(rdev_to_dev(rdev),
 +                       " on device, QoS may not be fully functional\n");
 +      }
 +      qcfgmap = &resp.pri0_cos_queue_id;
 +      tmp_map = (u8 *)cid_map;
 +      for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++)
 +              tmp_map[i] = qcfgmap[i];
 +
 +      return rc;
 +}
 +
 +static bool bnxt_re_is_qp1_or_shadow_qp(struct bnxt_re_dev *rdev,
 +                                      struct bnxt_re_qp *qp)
 +{
 +      return (qp->ib_qp.qp_type == IB_QPT_GSI) || (qp == rdev->qp1_sqp);
 +}
 +
 +static void bnxt_re_dev_stop(struct bnxt_re_dev *rdev)
 +{
 +      int mask = IB_QP_STATE;
 +      struct ib_qp_attr qp_attr;
 +      struct bnxt_re_qp *qp;
 +
 +      qp_attr.qp_state = IB_QPS_ERR;
 +      mutex_lock(&rdev->qp_lock);
 +      list_for_each_entry(qp, &rdev->qp_list, list) {
 +              /* Modify the state of all QPs except QP1/Shadow QP */
 +              if (!bnxt_re_is_qp1_or_shadow_qp(rdev, qp)) {
 +                      if (qp->qplib_qp.state !=
 +                          CMDQ_MODIFY_QP_NEW_STATE_RESET &&
 +                          qp->qplib_qp.state !=
 +                          CMDQ_MODIFY_QP_NEW_STATE_ERR) {
 +                              bnxt_re_dispatch_event(&rdev->ibdev, &qp->ib_qp,
 +                                                     1, IB_EVENT_QP_FATAL);
 +                              bnxt_re_modify_qp(&qp->ib_qp, &qp_attr, mask,
 +                                                NULL);
 +                      }
 +              }
 +      }
 +      mutex_unlock(&rdev->qp_lock);
 +}
 +
 +static u32 bnxt_re_get_priority_mask(struct bnxt_re_dev *rdev)
 +{
 +      u32 prio_map = 0, tmp_map = 0;
 +      struct net_device *netdev;
 +      struct dcb_app app;
 +
 +      netdev = rdev->netdev;
 +
 +      memset(&app, 0, sizeof(app));
 +      app.selector = IEEE_8021QAZ_APP_SEL_ETHERTYPE;
 +      app.protocol = ETH_P_IBOE;
 +      tmp_map = dcb_ieee_getapp_mask(netdev, &app);
 +      prio_map = tmp_map;
 +
 +      app.selector = IEEE_8021QAZ_APP_SEL_DGRAM;
 +      app.protocol = ROCE_V2_UDP_DPORT;
 +      tmp_map = dcb_ieee_getapp_mask(netdev, &app);
 +      prio_map |= tmp_map;
 +
 +      if (!prio_map)
 +              prio_map = -EFAULT;
 +      return prio_map;
 +}
 +
 +static void bnxt_re_parse_cid_map(u8 prio_map, u8 *cid_map, u16 *cosq)
 +{
 +      u16 prio;
 +      u8 id;
 +
 +      for (prio = 0, id = 0; prio < 8; prio++) {
 +              if (prio_map & (1 << prio)) {
 +                      cosq[id] = cid_map[prio];
 +                      id++;
 +                      if (id == 2) /* Max 2 tcs supported */
 +                              break;
 +              }
 +      }
 +}
 +
 +static int bnxt_re_setup_qos(struct bnxt_re_dev *rdev)
 +{
 +      u8 prio_map = 0;
 +      u64 cid_map;
 +      int rc;
 +
 +      /* Get priority for roce */
 +      rc = bnxt_re_get_priority_mask(rdev);
 +      if (rc < 0)
 +              return rc;
 +      prio_map = (u8)rc;
 +
 +      if (prio_map == rdev->cur_prio_map)
 +              return 0;
 +      rdev->cur_prio_map = prio_map;
 +      /* Get cosq id for this priority */
 +      rc = bnxt_re_query_hwrm_pri2cos(rdev, 0, &cid_map);
 +      if (rc) {
 +              dev_warn(rdev_to_dev(rdev), "no cos for p_mask %x\n", prio_map);
 +              return rc;
 +      }
 +      /* Parse CoS IDs for app priority */
 +      bnxt_re_parse_cid_map(prio_map, (u8 *)&cid_map, rdev->cosq);
 +
 +      /* Config BONO. */
 +      rc = bnxt_qplib_map_tc2cos(&rdev->qplib_res, rdev->cosq);
 +      if (rc) {
 +              dev_warn(rdev_to_dev(rdev), "no tc for cos{%x, %x}\n",
 +                       rdev->cosq[0], rdev->cosq[1]);
 +              return rc;
 +      }
 +
 +      return 0;
 +}
 +
 +static void bnxt_re_ib_unreg(struct bnxt_re_dev *rdev, bool lock_wait)
 +{
 +      int i, rc;
 +
 +      if (test_and_clear_bit(BNXT_RE_FLAG_IBDEV_REGISTERED, &rdev->flags)) {
 +              for (i = 0; i < ARRAY_SIZE(bnxt_re_attributes); i++)
 +                      device_remove_file(&rdev->ibdev.dev,
 +                                         bnxt_re_attributes[i]);
 +              /* Cleanup ib dev */
 +              bnxt_re_unregister_ib(rdev);
 +      }
 +      if (test_and_clear_bit(BNXT_RE_FLAG_QOS_WORK_REG, &rdev->flags))
 +              cancel_delayed_work(&rdev->worker);
 +
 +      bnxt_re_cleanup_res(rdev);
 +      bnxt_re_free_res(rdev, lock_wait);
 +
 +      if (test_and_clear_bit(BNXT_RE_FLAG_RCFW_CHANNEL_EN, &rdev->flags)) {
 +              rc = bnxt_qplib_deinit_rcfw(&rdev->rcfw);
 +              if (rc)
 +                      dev_warn(rdev_to_dev(rdev),
 +                               "Failed to deinitialize RCFW: %#x", rc);
 +              bnxt_re_net_stats_ctx_free(rdev, rdev->qplib_ctx.stats.fw_id,
 +                                         lock_wait);
 +              bnxt_qplib_free_ctx(rdev->en_dev->pdev, &rdev->qplib_ctx);
 +              bnxt_qplib_disable_rcfw_channel(&rdev->rcfw);
 +              bnxt_re_net_ring_free(rdev, rdev->rcfw.creq_ring_id, lock_wait);
 +              bnxt_qplib_free_rcfw_channel(&rdev->rcfw);
 +      }
 +      if (test_and_clear_bit(BNXT_RE_FLAG_GOT_MSIX, &rdev->flags)) {
 +              rc = bnxt_re_free_msix(rdev, lock_wait);
 +              if (rc)
 +                      dev_warn(rdev_to_dev(rdev),
 +                               "Failed to free MSI-X vectors: %#x", rc);
 +      }
 +      if (test_and_clear_bit(BNXT_RE_FLAG_NETDEV_REGISTERED, &rdev->flags)) {
 +              rc = bnxt_re_unregister_netdev(rdev, lock_wait);
 +              if (rc)
 +                      dev_warn(rdev_to_dev(rdev),
 +                               "Failed to unregister with netdev: %#x", rc);
 +      }
 +}
 +
 +static void bnxt_re_set_resource_limits(struct bnxt_re_dev *rdev)
 +{
 +      u32 i;
 +
 +      rdev->qplib_ctx.qpc_count = BNXT_RE_MAX_QPC_COUNT;
 +      rdev->qplib_ctx.mrw_count = BNXT_RE_MAX_MRW_COUNT;
 +      rdev->qplib_ctx.srqc_count = BNXT_RE_MAX_SRQC_COUNT;
 +      rdev->qplib_ctx.cq_count = BNXT_RE_MAX_CQ_COUNT;
 +      for (i = 0; i < MAX_TQM_ALLOC_REQ; i++)
 +              rdev->qplib_ctx.tqm_count[i] =
 +              rdev->dev_attr.tqm_alloc_reqs[i];
 +}
 +
 +/* worker thread for polling periodic events. Now used for QoS programming*/
 +static void bnxt_re_worker(struct work_struct *work)
 +{
 +      struct bnxt_re_dev *rdev = container_of(work, struct bnxt_re_dev,
 +                                              worker.work);
 +
 +      bnxt_re_setup_qos(rdev);
 +      schedule_delayed_work(&rdev->worker, msecs_to_jiffies(30000));
 +}
 +
 +static int bnxt_re_ib_reg(struct bnxt_re_dev *rdev)
 +{
 +      int i, j, rc;
 +
 +      /* Registered a new RoCE device instance to netdev */
 +      rc = bnxt_re_register_netdev(rdev);
 +      if (rc) {
 +              pr_err("Failed to register with netedev: %#x\n", rc);
 +              return -EINVAL;
 +      }
 +      set_bit(BNXT_RE_FLAG_NETDEV_REGISTERED, &rdev->flags);
 +
 +      rc = bnxt_re_request_msix(rdev);
 +      if (rc) {
 +              pr_err("Failed to get MSI-X vectors: %#x\n", rc);
 +              rc = -EINVAL;
 +              goto fail;
 +      }
 +      set_bit(BNXT_RE_FLAG_GOT_MSIX, &rdev->flags);
 +
 +      /* Establish RCFW Communication Channel to initialize the context
 +       * memory for the function and all child VFs
 +       */
 +      rc = bnxt_qplib_alloc_rcfw_channel(rdev->en_dev->pdev, &rdev->rcfw);
 +      if (rc)
 +              goto fail;
 +
 +      rc = bnxt_re_net_ring_alloc
 +                      (rdev, rdev->rcfw.creq.pbl[PBL_LVL_0].pg_map_arr,
 +                       rdev->rcfw.creq.pbl[rdev->rcfw.creq.level].pg_count,
 +                       HWRM_RING_ALLOC_CMPL, BNXT_QPLIB_CREQE_MAX_CNT - 1,
 +                       rdev->msix_entries[BNXT_RE_AEQ_IDX].ring_idx,
 +                       &rdev->rcfw.creq_ring_id);
 +      if (rc) {
 +              pr_err("Failed to allocate CREQ: %#x\n", rc);
 +              goto free_rcfw;
 +      }
 +      rc = bnxt_qplib_enable_rcfw_channel
 +                              (rdev->en_dev->pdev, &rdev->rcfw,
 +                               rdev->msix_entries[BNXT_RE_AEQ_IDX].vector,
 +                               rdev->msix_entries[BNXT_RE_AEQ_IDX].db_offset,
 +                               0, &bnxt_re_aeq_handler);
 +      if (rc) {
 +              pr_err("Failed to enable RCFW channel: %#x\n", rc);
 +              goto free_ring;
 +      }
 +
 +      rc = bnxt_qplib_get_dev_attr(&rdev->rcfw, &rdev->dev_attr);
 +      if (rc)
 +              goto disable_rcfw;
 +      bnxt_re_set_resource_limits(rdev);
 +
 +      rc = bnxt_qplib_alloc_ctx(rdev->en_dev->pdev, &rdev->qplib_ctx, 0);
 +      if (rc) {
 +              pr_err("Failed to allocate QPLIB context: %#x\n", rc);
 +              goto disable_rcfw;
 +      }
 +      rc = bnxt_re_net_stats_ctx_alloc(rdev,
 +                                       rdev->qplib_ctx.stats.dma_map,
 +                                       &rdev->qplib_ctx.stats.fw_id);
 +      if (rc) {
 +              pr_err("Failed to allocate stats context: %#x\n", rc);
 +              goto free_ctx;
 +      }
 +
 +      rc = bnxt_qplib_init_rcfw(&rdev->rcfw, &rdev->qplib_ctx, 0);
 +      if (rc) {
 +              pr_err("Failed to initialize RCFW: %#x\n", rc);
 +              goto free_sctx;
 +      }
 +      set_bit(BNXT_RE_FLAG_RCFW_CHANNEL_EN, &rdev->flags);
 +
 +      /* Resources based on the 'new' device caps */
 +      rc = bnxt_re_alloc_res(rdev);
 +      if (rc) {
 +              pr_err("Failed to allocate resources: %#x\n", rc);
 +              goto fail;
 +      }
 +      rc = bnxt_re_init_res(rdev);
 +      if (rc) {
 +              pr_err("Failed to initialize resources: %#x\n", rc);
 +              goto fail;
 +      }
 +
 +      rc = bnxt_re_setup_qos(rdev);
 +      if (rc)
 +              pr_info("RoCE priority not yet configured\n");
 +
 +      INIT_DELAYED_WORK(&rdev->worker, bnxt_re_worker);
 +      set_bit(BNXT_RE_FLAG_QOS_WORK_REG, &rdev->flags);
 +      schedule_delayed_work(&rdev->worker, msecs_to_jiffies(30000));
 +
 +      /* Register ib dev */
 +      rc = bnxt_re_register_ib(rdev);
 +      if (rc) {
 +              pr_err("Failed to register with IB: %#x\n", rc);
 +              goto fail;
 +      }
 +      dev_info(rdev_to_dev(rdev), "Device registered successfully");
 +      for (i = 0; i < ARRAY_SIZE(bnxt_re_attributes); i++) {
 +              rc = device_create_file(&rdev->ibdev.dev,
 +                                      bnxt_re_attributes[i]);
 +              if (rc) {
 +                      dev_err(rdev_to_dev(rdev),
 +                              "Failed to create IB sysfs: %#x", rc);
 +                      /* Must clean up all created device files */
 +                      for (j = 0; j < i; j++)
 +                              device_remove_file(&rdev->ibdev.dev,
 +                                                 bnxt_re_attributes[j]);
 +                      bnxt_re_unregister_ib(rdev);
 +                      goto fail;
 +              }
 +      }
 +      set_bit(BNXT_RE_FLAG_IBDEV_REGISTERED, &rdev->flags);
 +      bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1, IB_EVENT_PORT_ACTIVE);
 +      bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1, IB_EVENT_GID_CHANGE);
 +
 +      return 0;
 +free_sctx:
 +      bnxt_re_net_stats_ctx_free(rdev, rdev->qplib_ctx.stats.fw_id, true);
 +free_ctx:
 +      bnxt_qplib_free_ctx(rdev->en_dev->pdev, &rdev->qplib_ctx);
 +disable_rcfw:
 +      bnxt_qplib_disable_rcfw_channel(&rdev->rcfw);
 +free_ring:
 +      bnxt_re_net_ring_free(rdev, rdev->rcfw.creq_ring_id, true);
 +free_rcfw:
 +      bnxt_qplib_free_rcfw_channel(&rdev->rcfw);
 +fail:
 +      bnxt_re_ib_unreg(rdev, true);
 +      return rc;
 +}
 +
 +static void bnxt_re_dev_unreg(struct bnxt_re_dev *rdev)
 +{
 +      struct bnxt_en_dev *en_dev = rdev->en_dev;
 +      struct net_device *netdev = rdev->netdev;
 +
 +      bnxt_re_dev_remove(rdev);
 +
 +      if (netdev)
 +              bnxt_re_dev_unprobe(netdev, en_dev);
 +}
 +
 +static int bnxt_re_dev_reg(struct bnxt_re_dev **rdev, struct net_device *netdev)
 +{
 +      struct bnxt_en_dev *en_dev;
 +      int rc = 0;
 +
 +      if (!is_bnxt_re_dev(netdev))
 +              return -ENODEV;
 +
 +      en_dev = bnxt_re_dev_probe(netdev);
 +      if (IS_ERR(en_dev)) {
 +              if (en_dev != ERR_PTR(-ENODEV))
 +                      pr_err("%s: Failed to probe\n", ROCE_DRV_MODULE_NAME);
 +              rc = PTR_ERR(en_dev);
 +              goto exit;
 +      }
 +      *rdev = bnxt_re_dev_add(netdev, en_dev);
 +      if (!*rdev) {
 +              rc = -ENOMEM;
 +              bnxt_re_dev_unprobe(netdev, en_dev);
 +              goto exit;
 +      }
 +exit:
 +      return rc;
 +}
 +
 +static void bnxt_re_remove_one(struct bnxt_re_dev *rdev)
 +{
 +      pci_dev_put(rdev->en_dev->pdev);
 +}
 +
 +/* Handle all deferred netevents tasks */
 +static void bnxt_re_task(struct work_struct *work)
 +{
 +      struct bnxt_re_work *re_work;
 +      struct bnxt_re_dev *rdev;
 +      int rc = 0;
 +
 +      re_work = container_of(work, struct bnxt_re_work, work);
 +      rdev = re_work->rdev;
 +
 +      if (re_work->event != NETDEV_REGISTER &&
 +          !test_bit(BNXT_RE_FLAG_IBDEV_REGISTERED, &rdev->flags))
 +              return;
 +
 +      switch (re_work->event) {
 +      case NETDEV_REGISTER:
 +              rc = bnxt_re_ib_reg(rdev);
 +              if (rc)
 +                      dev_err(rdev_to_dev(rdev),
 +                              "Failed to register with IB: %#x", rc);
 +              break;
 +      case NETDEV_UP:
 +              bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1,
 +                                     IB_EVENT_PORT_ACTIVE);
 +              break;
 +      case NETDEV_DOWN:
 +              bnxt_re_dev_stop(rdev);
 +              break;
 +      case NETDEV_CHANGE:
 +              if (!netif_carrier_ok(rdev->netdev))
 +                      bnxt_re_dev_stop(rdev);
 +              else if (netif_carrier_ok(rdev->netdev))
 +                      bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1,
 +                                             IB_EVENT_PORT_ACTIVE);
 +              break;
 +      default:
 +              break;
 +      }
 +      kfree(re_work);
 +}
 +
 +static void bnxt_re_init_one(struct bnxt_re_dev *rdev)
 +{
 +      pci_dev_get(rdev->en_dev->pdev);
 +}
 +
 +/*
 + * "Notifier chain callback can be invoked for the same chain from
 + * different CPUs at the same time".
 + *
 + * For cases when the netdev is already present, our call to the
 + * register_netdevice_notifier() will actually get the rtnl_lock()
 + * before sending NETDEV_REGISTER and (if up) NETDEV_UP
 + * events.
 + *
 + * But for cases when the netdev is not already present, the notifier
 + * chain is subjected to be invoked from different CPUs simultaneously.
 + *
 + * This is protected by the netdev_mutex.
 + */
 +static int bnxt_re_netdev_event(struct notifier_block *notifier,
 +                              unsigned long event, void *ptr)
 +{
 +      struct net_device *real_dev, *netdev = netdev_notifier_info_to_dev(ptr);
 +      struct bnxt_re_work *re_work;
 +      struct bnxt_re_dev *rdev;
 +      int rc = 0;
 +      bool sch_work = false;
 +
 +      real_dev = rdma_vlan_dev_real_dev(netdev);
 +      if (!real_dev)
 +              real_dev = netdev;
 +
 +      rdev = bnxt_re_from_netdev(real_dev);
 +      if (!rdev && event != NETDEV_REGISTER)
 +              goto exit;
 +      if (real_dev != netdev)
 +              goto exit;
 +
 +      switch (event) {
 +      case NETDEV_REGISTER:
 +              if (rdev)
 +                      break;
 +              rc = bnxt_re_dev_reg(&rdev, real_dev);
 +              if (rc == -ENODEV)
 +                      break;
 +              if (rc) {
 +                      pr_err("Failed to register with the device %s: %#x\n",
 +                             real_dev->name, rc);
 +                      break;
 +              }
 +              bnxt_re_init_one(rdev);
 +              sch_work = true;
 +              break;
 +
 +      case NETDEV_UNREGISTER:
 +              bnxt_re_ib_unreg(rdev, false);
 +              bnxt_re_remove_one(rdev);
 +              bnxt_re_dev_unreg(rdev);
 +              break;
 +
 +      default:
 +              sch_work = true;
 +              break;
 +      }
 +      if (sch_work) {
 +              /* Allocate for the deferred task */
 +              re_work = kzalloc(sizeof(*re_work), GFP_ATOMIC);
 +              if (re_work) {
 +                      re_work->rdev = rdev;
 +                      re_work->event = event;
 +                      re_work->vlan_dev = (real_dev == netdev ?
 +                                           NULL : netdev);
 +                      INIT_WORK(&re_work->work, bnxt_re_task);
 +                      queue_work(bnxt_re_wq, &re_work->work);
 +              }
 +      }
 +
 +exit:
 +      return NOTIFY_DONE;
 +}
 +
 +static struct notifier_block bnxt_re_netdev_notifier = {
 +      .notifier_call = bnxt_re_netdev_event
 +};
 +
 +static int __init bnxt_re_mod_init(void)
 +{
 +      int rc = 0;
 +
 +      pr_info("%s: %s", ROCE_DRV_MODULE_NAME, version);
 +
 +      bnxt_re_wq = create_singlethread_workqueue("bnxt_re");
 +      if (!bnxt_re_wq)
 +              return -ENOMEM;
 +
 +      INIT_LIST_HEAD(&bnxt_re_dev_list);
 +
 +      rc = register_netdevice_notifier(&bnxt_re_netdev_notifier);
 +      if (rc) {
 +              pr_err("%s: Cannot register to netdevice_notifier",
 +                     ROCE_DRV_MODULE_NAME);
 +              goto err_netdev;
 +      }
 +      return 0;
 +
 +err_netdev:
 +      destroy_workqueue(bnxt_re_wq);
 +
 +      return rc;
 +}
 +
 +static void __exit bnxt_re_mod_exit(void)
 +{
 +      unregister_netdevice_notifier(&bnxt_re_netdev_notifier);
 +      if (bnxt_re_wq)
 +              destroy_workqueue(bnxt_re_wq);
 +}
 +
 +module_init(bnxt_re_mod_init);
 +module_exit(bnxt_re_mod_exit);
Simple merge
Simple merge
Simple merge
index 3c1f483d003f76d3330fcaa95091828dcb80898d,87ca81b85fd0f96285f08d3ef2e87bfaea3ff542..b8f9382a8b7dd72daf02b37804b912999dc500a1
@@@ -910,67 -924,64 +910,67 @@@ static struct mlx5_ib_mr *reg_umr(struc
        if (!mr)
                return ERR_PTR(-EAGAIN);
  
 -      err = dma_map_mr_pas(dev, umem, npages, page_shift, &size, &mr_pas,
 -                           &dma);
 -      if (err)
 -              goto free_mr;
 -
 -      mlx5_ib_init_umr_context(&umr_context);
 +      mr->ibmr.pd = pd;
 +      mr->umem = umem;
 +      mr->access_flags = access_flags;
 +      mr->desc_size = sizeof(struct mlx5_mtt);
 +      mr->mmkey.iova = virt_addr;
 +      mr->mmkey.size = len;
 +      mr->mmkey.pd = to_mpd(pd)->pdn;
  
 -      umrwr.wr.wr_cqe = &umr_context.cqe;
 -      prep_umr_reg_wqe(pd, &umrwr.wr, &sg, dma, npages, mr->mmkey.key,
 -                       page_shift, virt_addr, len, access_flags);
 +      err = mlx5_ib_update_xlt(mr, 0, npages, page_shift,
 +                               MLX5_IB_UPD_XLT_ENABLE);
  
 -      down(&umrc->sem);
 -      err = ib_post_send(umrc->qp, &umrwr.wr, &bad);
        if (err) {
 -              mlx5_ib_warn(dev, "post send failed, err %d\n", err);
 -              goto unmap_dma;
 -      } else {
 -              wait_for_completion(&umr_context.done);
 -              if (umr_context.status != IB_WC_SUCCESS) {
 -                      mlx5_ib_warn(dev, "reg umr failed\n");
 -                      err = -EFAULT;
 -              }
 +              mlx5_mr_cache_free(dev, mr);
 +              return ERR_PTR(err);
        }
  
 -      mr->mmkey.iova = virt_addr;
 -      mr->mmkey.size = len;
 -      mr->mmkey.pd = to_mpd(pd)->pdn;
 -
        mr->live = 1;
  
 -unmap_dma:
 -      up(&umrc->sem);
 -      dma_unmap_single(ddev, dma, size, DMA_TO_DEVICE);
 +      return mr;
 +}
 +
 +static inline int populate_xlt(struct mlx5_ib_mr *mr, int idx, int npages,
 +                             void *xlt, int page_shift, size_t size,
 +                             int flags)
 +{
 +      struct mlx5_ib_dev *dev = mr->dev;
 +      struct ib_umem *umem = mr->umem;
 +      if (flags & MLX5_IB_UPD_XLT_INDIRECT) {
 +              mlx5_odp_populate_klm(xlt, idx, npages, mr, flags);
 +              return npages;
 +      }
  
 -      kfree(mr_pas);
 +      npages = min_t(size_t, npages, ib_umem_num_pages(umem) - idx);
  
 -free_mr:
 -      if (err) {
 -              free_cached_mr(dev, mr);
 -              return ERR_PTR(err);
 +      if (!(flags & MLX5_IB_UPD_XLT_ZAP)) {
 +              __mlx5_ib_populate_pas(dev, umem, page_shift,
 +                                     idx, npages, xlt,
 +                                     MLX5_IB_MTT_PRESENT);
 +              /* Clear padding after the pages
 +               * brought from the umem.
 +               */
 +              memset(xlt + (npages * sizeof(struct mlx5_mtt)), 0,
 +                     size - npages * sizeof(struct mlx5_mtt));
        }
  
 -      return mr;
 +      return npages;
  }
  
 -#ifdef CONFIG_INFINIBAND_ON_DEMAND_PAGING
 -int mlx5_ib_update_mtt(struct mlx5_ib_mr *mr, u64 start_page_index, int npages,
 -                     int zap)
 +#define MLX5_MAX_UMR_CHUNK ((1 << (MLX5_MAX_UMR_SHIFT + 4)) - \
 +                          MLX5_UMR_MTT_ALIGNMENT)
 +#define MLX5_SPARE_UMR_CHUNK 0x10000
 +
 +int mlx5_ib_update_xlt(struct mlx5_ib_mr *mr, u64 idx, int npages,
 +                     int page_shift, int flags)
  {
        struct mlx5_ib_dev *dev = mr->dev;
-       struct device *ddev = dev->ib_dev.dma_device;
+       struct device *ddev = dev->ib_dev.dev.parent;
 -      struct umr_common *umrc = &dev->umrc;
 -      struct mlx5_ib_umr_context umr_context;
 -      struct ib_umem *umem = mr->umem;
 +      struct mlx5_ib_ucontext *uctx = NULL;
        int size;
 -      __be64 *pas;
 +      void *xlt;
        dma_addr_t dma;
 -      struct ib_send_wr *bad;
        struct mlx5_umr_wr wr;
        struct ib_sge sg;
        int err = 0;
Simple merge
Simple merge
Simple merge
index c33a4f84413cf8ce18cbbc188621778c882ba852,2a821d2fb5690909991101f5d67f283b16b756be..78b276a90401bf04b9a29523dc9dd1ff56a7da08
@@@ -7,7 -7,7 +7,7 @@@
  #
  obj-$(CONFIG_INFINIBAND_RDMAVT) += rdmavt.o
  
- rdmavt-y := vt.o ah.o cq.o dma.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
 -rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o srq.o \
 -      trace.o
++rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
 +      rc.o srq.o trace.o
  
  CFLAGS_trace.o = -I$(src)
Simple merge
Simple merge
Simple merge
index 30a6985909e0d95c446eb51abe17cf570dd18f73,da3caef1b7cb3b46a4bdce378b4614d631da08d2..5a887efb4bdf1b6405d3fc4342918084ae420b57
@@@ -651,8 -651,15 +651,8 @@@ iscsi_iser_session_create(struct iscsi_
                                                   SHOST_DIX_GUARD_CRC);
                }
  
 -              /*
 -               * Limit the sg_tablesize and max_sectors based on the device
 -               * max fastreg page list length.
 -               */
 -              shost->sg_tablesize = min_t(unsigned short, shost->sg_tablesize,
 -                      ib_conn->device->ib_device->attrs.max_fast_reg_page_list_len);
 -
                if (iscsi_host_add(shost,
-                                  ib_conn->device->ib_device->dma_device)) {
+                                  ib_conn->device->ib_device->dev.parent)) {
                        mutex_unlock(&iser_conn->state_mutex);
                        goto free_host;
                }
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
Simple merge
index 89f5bd4e1d5201c847ff77823b6b4159a741ef2f,f199c42b9a86f693bbe80f6c96a6055182548e26..d84849c5dc05397421d51776d68fdb09042388f8
@@@ -1840,56 -1778,12 +1840,9 @@@ struct ib_port_cache 
  struct ib_cache {
        rwlock_t                lock;
        struct ib_event_handler event_handler;
 -      struct ib_pkey_cache  **pkey_cache;
 -      struct ib_gid_table   **gid_cache;
 -      u8                     *lmc_cache;
 -      enum ib_port_state     *port_state_cache;
 +      struct ib_port_cache   *ports;
  };
  
- struct ib_dma_mapping_ops {
-       int             (*mapping_error)(struct ib_device *dev,
-                                        u64 dma_addr);
-       u64             (*map_single)(struct ib_device *dev,
-                                     void *ptr, size_t size,
-                                     enum dma_data_direction direction);
-       void            (*unmap_single)(struct ib_device *dev,
-                                       u64 addr, size_t size,
-                                       enum dma_data_direction direction);
-       u64             (*map_page)(struct ib_device *dev,
-                                   struct page *page, unsigned long offset,
-                                   size_t size,
-                                   enum dma_data_direction direction);
-       void            (*unmap_page)(struct ib_device *dev,
-                                     u64 addr, size_t size,
-                                     enum dma_data_direction direction);
-       int             (*map_sg)(struct ib_device *dev,
-                                 struct scatterlist *sg, int nents,
-                                 enum dma_data_direction direction);
-       void            (*unmap_sg)(struct ib_device *dev,
-                                   struct scatterlist *sg, int nents,
-                                   enum dma_data_direction direction);
-       int             (*map_sg_attrs)(struct ib_device *dev,
-                                       struct scatterlist *sg, int nents,
-                                       enum dma_data_direction direction,
-                                       unsigned long attrs);
-       void            (*unmap_sg_attrs)(struct ib_device *dev,
-                                         struct scatterlist *sg, int nents,
-                                         enum dma_data_direction direction,
-                                         unsigned long attrs);
-       void            (*sync_single_for_cpu)(struct ib_device *dev,
-                                              u64 dma_handle,
-                                              size_t size,
-                                              enum dma_data_direction dir);
-       void            (*sync_single_for_device)(struct ib_device *dev,
-                                                 u64 dma_handle,
-                                                 size_t size,
-                                                 enum dma_data_direction dir);
-       void            *(*alloc_coherent)(struct ib_device *dev,
-                                          size_t size,
-                                          u64 *dma_handle,
-                                          gfp_t flag);
-       void            (*free_coherent)(struct ib_device *dev,
-                                        size_t size, void *cpu_addr,
-                                        u64 dma_handle);
- };
  struct iw_cm_verbs;
  
  struct ib_port_immutable {
diff --cc lib/Kconfig
Simple merge
diff --cc lib/Makefile
Simple merge
diff --cc net/rds/ib.h
index 540458928f3c8eab4529b4072826b882563c941e,4a520826da232a7db7309596e18b02fbf66d8672..ec550626e221cb80fa5aaec7a00116dc3bd55c5d
@@@ -232,10 -221,9 +232,10 @@@ struct rds_ib_device 
        spinlock_t              spinlock;       /* protect the above */
        atomic_t                refcount;
        struct work_struct      free_work;
 +      int                     *vector_load;
  };
  
- #define ibdev_to_node(ibdev) dev_to_node(ibdev->dma_device)
+ #define ibdev_to_node(ibdev) dev_to_node((ibdev)->dev.parent)
  #define rdsibdev_to_node(rdsibdev) ibdev_to_node(rdsibdev->dev)
  
  /* bits for i_ack_flags */