RDMA/efa: Add support for dmabuf memory regions
Implement a dmabuf importer for the EFA driver. As ODP is not supported, the pinned dmabuf are used to prevent the move_notify callback from being called. Link: https://lore.kernel.org/r/20211012120903.96933-4-galpress@amazon.com Signed-off-by: Gal Pressman <galpress@amazon.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
This commit is contained in:
parent
1e4df4a21c
commit
66f4817b57
3 changed files with 101 additions and 31 deletions
|
@ -154,6 +154,10 @@ int efa_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
|
||||||
struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
u64 virt_addr, int access_flags,
|
u64 virt_addr, int access_flags,
|
||||||
struct ib_udata *udata);
|
struct ib_udata *udata);
|
||||||
|
struct ib_mr *efa_reg_user_mr_dmabuf(struct ib_pd *ibpd, u64 start,
|
||||||
|
u64 length, u64 virt_addr,
|
||||||
|
int fd, int access_flags,
|
||||||
|
struct ib_udata *udata);
|
||||||
int efa_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
|
int efa_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
|
||||||
int efa_get_port_immutable(struct ib_device *ibdev, u32 port_num,
|
int efa_get_port_immutable(struct ib_device *ibdev, u32 port_num,
|
||||||
struct ib_port_immutable *immutable);
|
struct ib_port_immutable *immutable);
|
||||||
|
|
|
@ -380,6 +380,7 @@ static const struct ib_device_ops efa_dev_ops = {
|
||||||
.query_port = efa_query_port,
|
.query_port = efa_query_port,
|
||||||
.query_qp = efa_query_qp,
|
.query_qp = efa_query_qp,
|
||||||
.reg_user_mr = efa_reg_mr,
|
.reg_user_mr = efa_reg_mr,
|
||||||
|
.reg_user_mr_dmabuf = efa_reg_user_mr_dmabuf,
|
||||||
|
|
||||||
INIT_RDMA_OBJ_SIZE(ib_ah, efa_ah, ibah),
|
INIT_RDMA_OBJ_SIZE(ib_ah, efa_ah, ibah),
|
||||||
INIT_RDMA_OBJ_SIZE(ib_cq, efa_cq, ibcq),
|
INIT_RDMA_OBJ_SIZE(ib_cq, efa_cq, ibcq),
|
||||||
|
|
|
@ -3,6 +3,8 @@
|
||||||
* Copyright 2018-2021 Amazon.com, Inc. or its affiliates. All rights reserved.
|
* Copyright 2018-2021 Amazon.com, Inc. or its affiliates. All rights reserved.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
#include <linux/dma-buf.h>
|
||||||
|
#include <linux/dma-resv.h>
|
||||||
#include <linux/vmalloc.h>
|
#include <linux/vmalloc.h>
|
||||||
#include <linux/log2.h>
|
#include <linux/log2.h>
|
||||||
|
|
||||||
|
@ -1544,26 +1546,18 @@ static int efa_create_pbl(struct efa_dev *dev,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
static struct efa_mr *efa_alloc_mr(struct ib_pd *ibpd, int access_flags,
|
||||||
u64 virt_addr, int access_flags,
|
struct ib_udata *udata)
|
||||||
struct ib_udata *udata)
|
|
||||||
{
|
{
|
||||||
struct efa_dev *dev = to_edev(ibpd->device);
|
struct efa_dev *dev = to_edev(ibpd->device);
|
||||||
struct efa_com_reg_mr_params params = {};
|
|
||||||
struct efa_com_reg_mr_result result = {};
|
|
||||||
struct pbl_context pbl;
|
|
||||||
int supp_access_flags;
|
int supp_access_flags;
|
||||||
unsigned int pg_sz;
|
|
||||||
struct efa_mr *mr;
|
struct efa_mr *mr;
|
||||||
int inline_size;
|
|
||||||
int err;
|
|
||||||
|
|
||||||
if (udata && udata->inlen &&
|
if (udata && udata->inlen &&
|
||||||
!ib_is_udata_cleared(udata, 0, sizeof(udata->inlen))) {
|
!ib_is_udata_cleared(udata, 0, sizeof(udata->inlen))) {
|
||||||
ibdev_dbg(&dev->ibdev,
|
ibdev_dbg(&dev->ibdev,
|
||||||
"Incompatible ABI params, udata not cleared\n");
|
"Incompatible ABI params, udata not cleared\n");
|
||||||
err = -EINVAL;
|
return ERR_PTR(-EINVAL);
|
||||||
goto err_out;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
supp_access_flags =
|
supp_access_flags =
|
||||||
|
@ -1575,23 +1569,26 @@ struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
ibdev_dbg(&dev->ibdev,
|
ibdev_dbg(&dev->ibdev,
|
||||||
"Unsupported access flags[%#x], supported[%#x]\n",
|
"Unsupported access flags[%#x], supported[%#x]\n",
|
||||||
access_flags, supp_access_flags);
|
access_flags, supp_access_flags);
|
||||||
err = -EOPNOTSUPP;
|
return ERR_PTR(-EOPNOTSUPP);
|
||||||
goto err_out;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
mr = kzalloc(sizeof(*mr), GFP_KERNEL);
|
mr = kzalloc(sizeof(*mr), GFP_KERNEL);
|
||||||
if (!mr) {
|
if (!mr)
|
||||||
err = -ENOMEM;
|
return ERR_PTR(-ENOMEM);
|
||||||
goto err_out;
|
|
||||||
}
|
|
||||||
|
|
||||||
mr->umem = ib_umem_get(ibpd->device, start, length, access_flags);
|
return mr;
|
||||||
if (IS_ERR(mr->umem)) {
|
}
|
||||||
err = PTR_ERR(mr->umem);
|
|
||||||
ibdev_dbg(&dev->ibdev,
|
static int efa_register_mr(struct ib_pd *ibpd, struct efa_mr *mr, u64 start,
|
||||||
"Failed to pin and map user space memory[%d]\n", err);
|
u64 length, u64 virt_addr, int access_flags)
|
||||||
goto err_free;
|
{
|
||||||
}
|
struct efa_dev *dev = to_edev(ibpd->device);
|
||||||
|
struct efa_com_reg_mr_params params = {};
|
||||||
|
struct efa_com_reg_mr_result result = {};
|
||||||
|
struct pbl_context pbl;
|
||||||
|
unsigned int pg_sz;
|
||||||
|
int inline_size;
|
||||||
|
int err;
|
||||||
|
|
||||||
params.pd = to_epd(ibpd)->pdn;
|
params.pd = to_epd(ibpd)->pdn;
|
||||||
params.iova = virt_addr;
|
params.iova = virt_addr;
|
||||||
|
@ -1602,10 +1599,9 @@ struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
dev->dev_attr.page_size_cap,
|
dev->dev_attr.page_size_cap,
|
||||||
virt_addr);
|
virt_addr);
|
||||||
if (!pg_sz) {
|
if (!pg_sz) {
|
||||||
err = -EOPNOTSUPP;
|
|
||||||
ibdev_dbg(&dev->ibdev, "Failed to find a suitable page size in page_size_cap %#llx\n",
|
ibdev_dbg(&dev->ibdev, "Failed to find a suitable page size in page_size_cap %#llx\n",
|
||||||
dev->dev_attr.page_size_cap);
|
dev->dev_attr.page_size_cap);
|
||||||
goto err_unmap;
|
return -EOPNOTSUPP;
|
||||||
}
|
}
|
||||||
|
|
||||||
params.page_shift = order_base_2(pg_sz);
|
params.page_shift = order_base_2(pg_sz);
|
||||||
|
@ -1619,21 +1615,21 @@ struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
if (params.page_num <= inline_size) {
|
if (params.page_num <= inline_size) {
|
||||||
err = efa_create_inline_pbl(dev, mr, ¶ms);
|
err = efa_create_inline_pbl(dev, mr, ¶ms);
|
||||||
if (err)
|
if (err)
|
||||||
goto err_unmap;
|
return err;
|
||||||
|
|
||||||
err = efa_com_register_mr(&dev->edev, ¶ms, &result);
|
err = efa_com_register_mr(&dev->edev, ¶ms, &result);
|
||||||
if (err)
|
if (err)
|
||||||
goto err_unmap;
|
return err;
|
||||||
} else {
|
} else {
|
||||||
err = efa_create_pbl(dev, &pbl, mr, ¶ms);
|
err = efa_create_pbl(dev, &pbl, mr, ¶ms);
|
||||||
if (err)
|
if (err)
|
||||||
goto err_unmap;
|
return err;
|
||||||
|
|
||||||
err = efa_com_register_mr(&dev->edev, ¶ms, &result);
|
err = efa_com_register_mr(&dev->edev, ¶ms, &result);
|
||||||
pbl_destroy(dev, &pbl);
|
pbl_destroy(dev, &pbl);
|
||||||
|
|
||||||
if (err)
|
if (err)
|
||||||
goto err_unmap;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
mr->ibmr.lkey = result.l_key;
|
mr->ibmr.lkey = result.l_key;
|
||||||
|
@ -1641,9 +1637,78 @@ struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
mr->ibmr.length = length;
|
mr->ibmr.length = length;
|
||||||
ibdev_dbg(&dev->ibdev, "Registered mr[%d]\n", mr->ibmr.lkey);
|
ibdev_dbg(&dev->ibdev, "Registered mr[%d]\n", mr->ibmr.lkey);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct ib_mr *efa_reg_user_mr_dmabuf(struct ib_pd *ibpd, u64 start,
|
||||||
|
u64 length, u64 virt_addr,
|
||||||
|
int fd, int access_flags,
|
||||||
|
struct ib_udata *udata)
|
||||||
|
{
|
||||||
|
struct efa_dev *dev = to_edev(ibpd->device);
|
||||||
|
struct ib_umem_dmabuf *umem_dmabuf;
|
||||||
|
struct efa_mr *mr;
|
||||||
|
int err;
|
||||||
|
|
||||||
|
mr = efa_alloc_mr(ibpd, access_flags, udata);
|
||||||
|
if (IS_ERR(mr)) {
|
||||||
|
err = PTR_ERR(mr);
|
||||||
|
goto err_out;
|
||||||
|
}
|
||||||
|
|
||||||
|
umem_dmabuf = ib_umem_dmabuf_get_pinned(ibpd->device, start, length, fd,
|
||||||
|
access_flags);
|
||||||
|
if (IS_ERR(umem_dmabuf)) {
|
||||||
|
err = PTR_ERR(umem_dmabuf);
|
||||||
|
ibdev_dbg(&dev->ibdev, "Failed to get dmabuf umem[%d]\n", err);
|
||||||
|
goto err_free;
|
||||||
|
}
|
||||||
|
|
||||||
|
mr->umem = &umem_dmabuf->umem;
|
||||||
|
err = efa_register_mr(ibpd, mr, start, length, virt_addr, access_flags);
|
||||||
|
if (err)
|
||||||
|
goto err_release;
|
||||||
|
|
||||||
return &mr->ibmr;
|
return &mr->ibmr;
|
||||||
|
|
||||||
err_unmap:
|
err_release:
|
||||||
|
ib_umem_release(mr->umem);
|
||||||
|
err_free:
|
||||||
|
kfree(mr);
|
||||||
|
err_out:
|
||||||
|
atomic64_inc(&dev->stats.reg_mr_err);
|
||||||
|
return ERR_PTR(err);
|
||||||
|
}
|
||||||
|
|
||||||
|
struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length,
|
||||||
|
u64 virt_addr, int access_flags,
|
||||||
|
struct ib_udata *udata)
|
||||||
|
{
|
||||||
|
struct efa_dev *dev = to_edev(ibpd->device);
|
||||||
|
struct efa_mr *mr;
|
||||||
|
int err;
|
||||||
|
|
||||||
|
mr = efa_alloc_mr(ibpd, access_flags, udata);
|
||||||
|
if (IS_ERR(mr)) {
|
||||||
|
err = PTR_ERR(mr);
|
||||||
|
goto err_out;
|
||||||
|
}
|
||||||
|
|
||||||
|
mr->umem = ib_umem_get(ibpd->device, start, length, access_flags);
|
||||||
|
if (IS_ERR(mr->umem)) {
|
||||||
|
err = PTR_ERR(mr->umem);
|
||||||
|
ibdev_dbg(&dev->ibdev,
|
||||||
|
"Failed to pin and map user space memory[%d]\n", err);
|
||||||
|
goto err_free;
|
||||||
|
}
|
||||||
|
|
||||||
|
err = efa_register_mr(ibpd, mr, start, length, virt_addr, access_flags);
|
||||||
|
if (err)
|
||||||
|
goto err_release;
|
||||||
|
|
||||||
|
return &mr->ibmr;
|
||||||
|
|
||||||
|
err_release:
|
||||||
ib_umem_release(mr->umem);
|
ib_umem_release(mr->umem);
|
||||||
err_free:
|
err_free:
|
||||||
kfree(mr);
|
kfree(mr);
|
||||||
|
|
Loading…
Add table
Reference in a new issue