The commitf959dcd6dd
("dma-direct: Fix potential NULL pointer dereference") made dma_mask as mandetory field to be setup even for dma_virt_ops based dma devices. The commit in the fixes tag omitted setting up the dma_mask on virtual devices triggering the below trace when they were combined during the merge window. Fix it by setting empty DMA MASK for software based RDMA devices. WARNING: CPU: 1 PID: 8488 at kernel/dma/mapping.c:149 dma_map_page_attrs+0x493/0x700 CPU: 1 PID: 8488 Comm: syz-executor144 Not tainted 5.9.0-syzkaller #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011 RIP: 0010:dma_map_page_attrs+0x493/0x700 kernel/dma/mapping.c:149 Trace: dma_map_single_attrs include/linux/dma-mapping.h:279 [inline] ib_dma_map_single include/rdma/ib_verbs.h:3967 [inline] ib_mad_post_receive_mads+0x23f/0xd60 drivers/infiniband/core/mad.c:2715 ib_mad_port_start drivers/infiniband/core/mad.c:2862 [inline] ib_mad_port_open drivers/infiniband/core/mad.c:3016 [inline] ib_mad_init_device+0x72b/0x1400 drivers/infiniband/core/mad.c:3092 add_client_context+0x405/0x5e0 drivers/infiniband/core/device.c:680 enable_device_and_get+0x1d5/0x3c0 drivers/infiniband/core/device.c:1301 ib_register_device drivers/infiniband/core/device.c:1376 [inline] ib_register_device+0x7a7/0xa40 drivers/infiniband/core/device.c:1335 rxe_register_device+0x46d/0x570 drivers/infiniband/sw/rxe/rxe_verbs.c:1182 rxe_add+0x12fe/0x16d0 drivers/infiniband/sw/rxe/rxe.c:247 rxe_net_add+0x8c/0xe0 drivers/infiniband/sw/rxe/rxe_net.c:507 rxe_newlink drivers/infiniband/sw/rxe/rxe.c:269 [inline] rxe_newlink+0xb7/0xe0 drivers/infiniband/sw/rxe/rxe.c:250 nldev_newlink+0x30e/0x540 drivers/infiniband/core/nldev.c:1555 rdma_nl_rcv_msg+0x367/0x690 drivers/infiniband/core/netlink.c:195 rdma_nl_rcv_skb drivers/infiniband/core/netlink.c:239 [inline] rdma_nl_rcv+0x2f2/0x440 drivers/infiniband/core/netlink.c:259 netlink_unicast_kernel net/netlink/af_netlink.c:1304 [inline] netlink_unicast+0x533/0x7d0 net/netlink/af_netlink.c:1330 netlink_sendmsg+0x856/0xd90 net/netlink/af_netlink.c:1919 sock_sendmsg_nosec net/socket.c:651 [inline] sock_sendmsg+0xcf/0x120 net/socket.c:671 ____sys_sendmsg+0x6e8/0x810 net/socket.c:2353 ___sys_sendmsg+0xf3/0x170 net/socket.c:2407 __sys_sendmsg+0xe5/0x1b0 net/socket.c:2440 do_syscall_64+0x2d/0x70 arch/x86/entry/common.c:46 entry_SYSCALL_64_after_hwframe+0x44/0xa9 RIP: 0033:0x443699 Link: https://lore.kernel.org/r/20201030093803.278830-1-parav@nvidia.com Reported-by: syzbot+34dc2fea3478e659af01@syzkaller.appspotmail.com Fixes:e0477b34d9
("RDMA: Explicitly pass in the dma_device to ib_register_device") Signed-off-by: Parav Pandit <parav@nvidia.com> Tested-by: Guoqing Jiang <guoqing.jiang@cloud.ionos.com> Tested-by: Dennis Dalessandro <dennis.dalessandro@cornelisnetworks.com> Reviewed-by: Christoph Hellwig <hch@lst.de> Acked-by: Zhu Yanjun <yanjunz@nvidia.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
698 lines
18 KiB
C
698 lines
18 KiB
C
/*
|
|
* Copyright(c) 2016 - 2018 Intel Corporation.
|
|
*
|
|
* This file is provided under a dual BSD/GPLv2 license. When using or
|
|
* redistributing this file, you may do so under either license.
|
|
*
|
|
* GPL LICENSE SUMMARY
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of version 2 of the GNU General Public License as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* BSD LICENSE
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* - Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* - Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* - Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include "vt.h"
|
|
#include "cq.h"
|
|
#include "trace.h"
|
|
|
|
#define RVT_UVERBS_ABI_VERSION 2
|
|
|
|
MODULE_LICENSE("Dual BSD/GPL");
|
|
MODULE_DESCRIPTION("RDMA Verbs Transport Library");
|
|
|
|
static int rvt_init(void)
|
|
{
|
|
int ret = rvt_driver_cq_init();
|
|
|
|
if (ret)
|
|
pr_err("Error in driver CQ init.\n");
|
|
|
|
return ret;
|
|
}
|
|
module_init(rvt_init);
|
|
|
|
static void rvt_cleanup(void)
|
|
{
|
|
rvt_cq_exit();
|
|
}
|
|
module_exit(rvt_cleanup);
|
|
|
|
/**
|
|
* rvt_alloc_device - allocate rdi
|
|
* @size: how big of a structure to allocate
|
|
* @nports: number of ports to allocate array slots for
|
|
*
|
|
* Use IB core device alloc to allocate space for the rdi which is assumed to be
|
|
* inside of the ib_device. Any extra space that drivers require should be
|
|
* included in size.
|
|
*
|
|
* We also allocate a port array based on the number of ports.
|
|
*
|
|
* Return: pointer to allocated rdi
|
|
*/
|
|
struct rvt_dev_info *rvt_alloc_device(size_t size, int nports)
|
|
{
|
|
struct rvt_dev_info *rdi;
|
|
|
|
rdi = container_of(_ib_alloc_device(size), struct rvt_dev_info, ibdev);
|
|
if (!rdi)
|
|
return rdi;
|
|
|
|
rdi->ports = kcalloc(nports, sizeof(*rdi->ports), GFP_KERNEL);
|
|
if (!rdi->ports)
|
|
ib_dealloc_device(&rdi->ibdev);
|
|
|
|
return rdi;
|
|
}
|
|
EXPORT_SYMBOL(rvt_alloc_device);
|
|
|
|
/**
|
|
* rvt_dealloc_device - deallocate rdi
|
|
* @rdi: structure to free
|
|
*
|
|
* Free a structure allocated with rvt_alloc_device()
|
|
*/
|
|
void rvt_dealloc_device(struct rvt_dev_info *rdi)
|
|
{
|
|
kfree(rdi->ports);
|
|
ib_dealloc_device(&rdi->ibdev);
|
|
}
|
|
EXPORT_SYMBOL(rvt_dealloc_device);
|
|
|
|
static int rvt_query_device(struct ib_device *ibdev,
|
|
struct ib_device_attr *props,
|
|
struct ib_udata *uhw)
|
|
{
|
|
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
|
|
|
|
if (uhw->inlen || uhw->outlen)
|
|
return -EINVAL;
|
|
/*
|
|
* Return rvt_dev_info.dparms.props contents
|
|
*/
|
|
*props = rdi->dparms.props;
|
|
return 0;
|
|
}
|
|
|
|
static int rvt_modify_device(struct ib_device *device,
|
|
int device_modify_mask,
|
|
struct ib_device_modify *device_modify)
|
|
{
|
|
/*
|
|
* There is currently no need to supply this based on qib and hfi1.
|
|
* Future drivers may need to implement this though.
|
|
*/
|
|
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
/**
|
|
* rvt_query_port: Passes the query port call to the driver
|
|
* @ibdev: Verbs IB dev
|
|
* @port_num: port number, 1 based from ib core
|
|
* @props: structure to hold returned properties
|
|
*
|
|
* Return: 0 on success
|
|
*/
|
|
static int rvt_query_port(struct ib_device *ibdev, u8 port_num,
|
|
struct ib_port_attr *props)
|
|
{
|
|
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
|
|
struct rvt_ibport *rvp;
|
|
int port_index = ibport_num_to_idx(ibdev, port_num);
|
|
|
|
if (port_index < 0)
|
|
return -EINVAL;
|
|
|
|
rvp = rdi->ports[port_index];
|
|
/* props being zeroed by the caller, avoid zeroing it here */
|
|
props->sm_lid = rvp->sm_lid;
|
|
props->sm_sl = rvp->sm_sl;
|
|
props->port_cap_flags = rvp->port_cap_flags;
|
|
props->max_msg_sz = 0x80000000;
|
|
props->pkey_tbl_len = rvt_get_npkeys(rdi);
|
|
props->bad_pkey_cntr = rvp->pkey_violations;
|
|
props->qkey_viol_cntr = rvp->qkey_violations;
|
|
props->subnet_timeout = rvp->subnet_timeout;
|
|
props->init_type_reply = 0;
|
|
|
|
/* Populate the remaining ib_port_attr elements */
|
|
return rdi->driver_f.query_port_state(rdi, port_num, props);
|
|
}
|
|
|
|
/**
|
|
* rvt_modify_port
|
|
* @ibdev: Verbs IB dev
|
|
* @port_num: Port number, 1 based from ib core
|
|
* @port_modify_mask: How to change the port
|
|
* @props: Structure to fill in
|
|
*
|
|
* Return: 0 on success
|
|
*/
|
|
static int rvt_modify_port(struct ib_device *ibdev, u8 port_num,
|
|
int port_modify_mask, struct ib_port_modify *props)
|
|
{
|
|
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
|
|
struct rvt_ibport *rvp;
|
|
int ret = 0;
|
|
int port_index = ibport_num_to_idx(ibdev, port_num);
|
|
|
|
if (port_index < 0)
|
|
return -EINVAL;
|
|
|
|
rvp = rdi->ports[port_index];
|
|
if (port_modify_mask & IB_PORT_OPA_MASK_CHG) {
|
|
rvp->port_cap3_flags |= props->set_port_cap_mask;
|
|
rvp->port_cap3_flags &= ~props->clr_port_cap_mask;
|
|
} else {
|
|
rvp->port_cap_flags |= props->set_port_cap_mask;
|
|
rvp->port_cap_flags &= ~props->clr_port_cap_mask;
|
|
}
|
|
|
|
if (props->set_port_cap_mask || props->clr_port_cap_mask)
|
|
rdi->driver_f.cap_mask_chg(rdi, port_num);
|
|
if (port_modify_mask & IB_PORT_SHUTDOWN)
|
|
ret = rdi->driver_f.shut_down_port(rdi, port_num);
|
|
if (port_modify_mask & IB_PORT_RESET_QKEY_CNTR)
|
|
rvp->qkey_violations = 0;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* rvt_query_pkey - Return a pkey from the table at a given index
|
|
* @ibdev: Verbs IB dev
|
|
* @port_num: Port number, 1 based from ib core
|
|
* @index: Index into pkey table
|
|
* @pkey: returned pkey from the port pkey table
|
|
*
|
|
* Return: 0 on failure pkey otherwise
|
|
*/
|
|
static int rvt_query_pkey(struct ib_device *ibdev, u8 port_num, u16 index,
|
|
u16 *pkey)
|
|
{
|
|
/*
|
|
* Driver will be responsible for keeping rvt_dev_info.pkey_table up to
|
|
* date. This function will just return that value. There is no need to
|
|
* lock, if a stale value is read and sent to the user so be it there is
|
|
* no way to protect against that anyway.
|
|
*/
|
|
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
|
|
int port_index;
|
|
|
|
port_index = ibport_num_to_idx(ibdev, port_num);
|
|
if (port_index < 0)
|
|
return -EINVAL;
|
|
|
|
if (index >= rvt_get_npkeys(rdi))
|
|
return -EINVAL;
|
|
|
|
*pkey = rvt_get_pkey(rdi, port_index, index);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* rvt_query_gid - Return a gid from the table
|
|
* @ibdev: Verbs IB dev
|
|
* @port_num: Port number, 1 based from ib core
|
|
* @guid_index: Index in table
|
|
* @gid: Gid to return
|
|
*
|
|
* Return: 0 on success
|
|
*/
|
|
static int rvt_query_gid(struct ib_device *ibdev, u8 port_num,
|
|
int guid_index, union ib_gid *gid)
|
|
{
|
|
struct rvt_dev_info *rdi;
|
|
struct rvt_ibport *rvp;
|
|
int port_index;
|
|
|
|
/*
|
|
* Driver is responsible for updating the guid table. Which will be used
|
|
* to craft the return value. This will work similar to how query_pkey()
|
|
* is being done.
|
|
*/
|
|
port_index = ibport_num_to_idx(ibdev, port_num);
|
|
if (port_index < 0)
|
|
return -EINVAL;
|
|
|
|
rdi = ib_to_rvt(ibdev);
|
|
rvp = rdi->ports[port_index];
|
|
|
|
gid->global.subnet_prefix = rvp->gid_prefix;
|
|
|
|
return rdi->driver_f.get_guid_be(rdi, rvp, guid_index,
|
|
&gid->global.interface_id);
|
|
}
|
|
|
|
/**
|
|
* rvt_alloc_ucontext - Allocate a user context
|
|
* @uctx: Verbs context
|
|
* @udata: User data allocated
|
|
*/
|
|
static int rvt_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* rvt_dealloc_ucontext - Free a user context
|
|
* @context - Free this
|
|
*/
|
|
static void rvt_dealloc_ucontext(struct ib_ucontext *context)
|
|
{
|
|
return;
|
|
}
|
|
|
|
static int rvt_get_port_immutable(struct ib_device *ibdev, u8 port_num,
|
|
struct ib_port_immutable *immutable)
|
|
{
|
|
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
|
|
struct ib_port_attr attr;
|
|
int err, port_index;
|
|
|
|
port_index = ibport_num_to_idx(ibdev, port_num);
|
|
if (port_index < 0)
|
|
return -EINVAL;
|
|
|
|
immutable->core_cap_flags = rdi->dparms.core_cap_flags;
|
|
|
|
err = ib_query_port(ibdev, port_num, &attr);
|
|
if (err)
|
|
return err;
|
|
|
|
immutable->pkey_tbl_len = attr.pkey_tbl_len;
|
|
immutable->gid_tbl_len = attr.gid_tbl_len;
|
|
immutable->max_mad_size = rdi->dparms.max_mad_size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
enum {
|
|
MISC,
|
|
QUERY_DEVICE,
|
|
MODIFY_DEVICE,
|
|
QUERY_PORT,
|
|
MODIFY_PORT,
|
|
QUERY_PKEY,
|
|
QUERY_GID,
|
|
ALLOC_UCONTEXT,
|
|
DEALLOC_UCONTEXT,
|
|
GET_PORT_IMMUTABLE,
|
|
CREATE_QP,
|
|
MODIFY_QP,
|
|
DESTROY_QP,
|
|
QUERY_QP,
|
|
POST_SEND,
|
|
POST_RECV,
|
|
POST_SRQ_RECV,
|
|
CREATE_AH,
|
|
DESTROY_AH,
|
|
MODIFY_AH,
|
|
QUERY_AH,
|
|
CREATE_SRQ,
|
|
MODIFY_SRQ,
|
|
DESTROY_SRQ,
|
|
QUERY_SRQ,
|
|
ATTACH_MCAST,
|
|
DETACH_MCAST,
|
|
GET_DMA_MR,
|
|
REG_USER_MR,
|
|
DEREG_MR,
|
|
ALLOC_MR,
|
|
MAP_MR_SG,
|
|
ALLOC_FMR,
|
|
MAP_PHYS_FMR,
|
|
UNMAP_FMR,
|
|
DEALLOC_FMR,
|
|
MMAP,
|
|
CREATE_CQ,
|
|
DESTROY_CQ,
|
|
POLL_CQ,
|
|
REQ_NOTFIY_CQ,
|
|
RESIZE_CQ,
|
|
ALLOC_PD,
|
|
DEALLOC_PD,
|
|
_VERB_IDX_MAX /* Must always be last! */
|
|
};
|
|
|
|
static const struct ib_device_ops rvt_dev_ops = {
|
|
.uverbs_abi_ver = RVT_UVERBS_ABI_VERSION,
|
|
|
|
.alloc_mr = rvt_alloc_mr,
|
|
.alloc_pd = rvt_alloc_pd,
|
|
.alloc_ucontext = rvt_alloc_ucontext,
|
|
.attach_mcast = rvt_attach_mcast,
|
|
.create_ah = rvt_create_ah,
|
|
.create_cq = rvt_create_cq,
|
|
.create_qp = rvt_create_qp,
|
|
.create_srq = rvt_create_srq,
|
|
.dealloc_pd = rvt_dealloc_pd,
|
|
.dealloc_ucontext = rvt_dealloc_ucontext,
|
|
.dereg_mr = rvt_dereg_mr,
|
|
.destroy_ah = rvt_destroy_ah,
|
|
.destroy_cq = rvt_destroy_cq,
|
|
.destroy_qp = rvt_destroy_qp,
|
|
.destroy_srq = rvt_destroy_srq,
|
|
.detach_mcast = rvt_detach_mcast,
|
|
.get_dma_mr = rvt_get_dma_mr,
|
|
.get_port_immutable = rvt_get_port_immutable,
|
|
.map_mr_sg = rvt_map_mr_sg,
|
|
.mmap = rvt_mmap,
|
|
.modify_ah = rvt_modify_ah,
|
|
.modify_device = rvt_modify_device,
|
|
.modify_port = rvt_modify_port,
|
|
.modify_qp = rvt_modify_qp,
|
|
.modify_srq = rvt_modify_srq,
|
|
.poll_cq = rvt_poll_cq,
|
|
.post_recv = rvt_post_recv,
|
|
.post_send = rvt_post_send,
|
|
.post_srq_recv = rvt_post_srq_recv,
|
|
.query_ah = rvt_query_ah,
|
|
.query_device = rvt_query_device,
|
|
.query_gid = rvt_query_gid,
|
|
.query_pkey = rvt_query_pkey,
|
|
.query_port = rvt_query_port,
|
|
.query_qp = rvt_query_qp,
|
|
.query_srq = rvt_query_srq,
|
|
.reg_user_mr = rvt_reg_user_mr,
|
|
.req_notify_cq = rvt_req_notify_cq,
|
|
.resize_cq = rvt_resize_cq,
|
|
|
|
INIT_RDMA_OBJ_SIZE(ib_ah, rvt_ah, ibah),
|
|
INIT_RDMA_OBJ_SIZE(ib_cq, rvt_cq, ibcq),
|
|
INIT_RDMA_OBJ_SIZE(ib_pd, rvt_pd, ibpd),
|
|
INIT_RDMA_OBJ_SIZE(ib_srq, rvt_srq, ibsrq),
|
|
INIT_RDMA_OBJ_SIZE(ib_ucontext, rvt_ucontext, ibucontext),
|
|
};
|
|
|
|
static noinline int check_support(struct rvt_dev_info *rdi, int verb)
|
|
{
|
|
switch (verb) {
|
|
case MISC:
|
|
/*
|
|
* These functions are not part of verbs specifically but are
|
|
* required for rdmavt to function.
|
|
*/
|
|
if ((!rdi->ibdev.ops.init_port) ||
|
|
(!rdi->driver_f.get_pci_dev))
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case MODIFY_DEVICE:
|
|
/*
|
|
* rdmavt does not support modify device currently drivers must
|
|
* provide.
|
|
*/
|
|
if (!rdi->ibdev.ops.modify_device)
|
|
return -EOPNOTSUPP;
|
|
break;
|
|
|
|
case QUERY_PORT:
|
|
if (!rdi->ibdev.ops.query_port)
|
|
if (!rdi->driver_f.query_port_state)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case MODIFY_PORT:
|
|
if (!rdi->ibdev.ops.modify_port)
|
|
if (!rdi->driver_f.cap_mask_chg ||
|
|
!rdi->driver_f.shut_down_port)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case QUERY_GID:
|
|
if (!rdi->ibdev.ops.query_gid)
|
|
if (!rdi->driver_f.get_guid_be)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case CREATE_QP:
|
|
if (!rdi->ibdev.ops.create_qp)
|
|
if (!rdi->driver_f.qp_priv_alloc ||
|
|
!rdi->driver_f.qp_priv_free ||
|
|
!rdi->driver_f.notify_qp_reset ||
|
|
!rdi->driver_f.flush_qp_waiters ||
|
|
!rdi->driver_f.stop_send_queue ||
|
|
!rdi->driver_f.quiesce_qp)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case MODIFY_QP:
|
|
if (!rdi->ibdev.ops.modify_qp)
|
|
if (!rdi->driver_f.notify_qp_reset ||
|
|
!rdi->driver_f.schedule_send ||
|
|
!rdi->driver_f.get_pmtu_from_attr ||
|
|
!rdi->driver_f.flush_qp_waiters ||
|
|
!rdi->driver_f.stop_send_queue ||
|
|
!rdi->driver_f.quiesce_qp ||
|
|
!rdi->driver_f.notify_error_qp ||
|
|
!rdi->driver_f.mtu_from_qp ||
|
|
!rdi->driver_f.mtu_to_path_mtu)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case DESTROY_QP:
|
|
if (!rdi->ibdev.ops.destroy_qp)
|
|
if (!rdi->driver_f.qp_priv_free ||
|
|
!rdi->driver_f.notify_qp_reset ||
|
|
!rdi->driver_f.flush_qp_waiters ||
|
|
!rdi->driver_f.stop_send_queue ||
|
|
!rdi->driver_f.quiesce_qp)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
case POST_SEND:
|
|
if (!rdi->ibdev.ops.post_send)
|
|
if (!rdi->driver_f.schedule_send ||
|
|
!rdi->driver_f.do_send ||
|
|
!rdi->post_parms)
|
|
return -EINVAL;
|
|
break;
|
|
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* rvt_register_device - register a driver
|
|
* @rdi: main dev structure for all of rdmavt operations
|
|
*
|
|
* It is up to drivers to allocate the rdi and fill in the appropriate
|
|
* information.
|
|
*
|
|
* Return: 0 on success otherwise an errno.
|
|
*/
|
|
int rvt_register_device(struct rvt_dev_info *rdi)
|
|
{
|
|
int ret = 0, i;
|
|
u64 dma_mask;
|
|
|
|
if (!rdi)
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* Check to ensure drivers have setup the required helpers for the verbs
|
|
* they want rdmavt to handle
|
|
*/
|
|
for (i = 0; i < _VERB_IDX_MAX; i++)
|
|
if (check_support(rdi, i)) {
|
|
pr_err("Driver support req not met at %d\n", i);
|
|
return -EINVAL;
|
|
}
|
|
|
|
ib_set_device_ops(&rdi->ibdev, &rvt_dev_ops);
|
|
|
|
/* Once we get past here we can use rvt_pr macros and tracepoints */
|
|
trace_rvt_dbg(rdi, "Driver attempting registration");
|
|
rvt_mmap_init(rdi);
|
|
|
|
/* Queue Pairs */
|
|
ret = rvt_driver_qp_init(rdi);
|
|
if (ret) {
|
|
pr_err("Error in driver QP init.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Address Handle */
|
|
spin_lock_init(&rdi->n_ahs_lock);
|
|
rdi->n_ahs_allocated = 0;
|
|
|
|
/* Shared Receive Queue */
|
|
rvt_driver_srq_init(rdi);
|
|
|
|
/* Multicast */
|
|
rvt_driver_mcast_init(rdi);
|
|
|
|
/* Mem Region */
|
|
ret = rvt_driver_mr_init(rdi);
|
|
if (ret) {
|
|
pr_err("Error in driver MR init.\n");
|
|
goto bail_no_mr;
|
|
}
|
|
|
|
/* Memory Working Set Size */
|
|
ret = rvt_wss_init(rdi);
|
|
if (ret) {
|
|
rvt_pr_err(rdi, "Error in WSS init.\n");
|
|
goto bail_mr;
|
|
}
|
|
|
|
/* Completion queues */
|
|
spin_lock_init(&rdi->n_cqs_lock);
|
|
|
|
/* DMA Operations */
|
|
rdi->ibdev.dev.dma_parms = rdi->ibdev.dev.parent->dma_parms;
|
|
dma_mask = IS_ENABLED(CONFIG_64BIT) ? DMA_BIT_MASK(64) : DMA_BIT_MASK(32);
|
|
ret = dma_coerce_mask_and_coherent(&rdi->ibdev.dev, dma_mask);
|
|
if (ret)
|
|
goto bail_wss;
|
|
|
|
/* Protection Domain */
|
|
spin_lock_init(&rdi->n_pds_lock);
|
|
rdi->n_pds_allocated = 0;
|
|
|
|
/*
|
|
* There are some things which could be set by underlying drivers but
|
|
* really should be up to rdmavt to set. For instance drivers can't know
|
|
* exactly which functions rdmavt supports, nor do they know the ABI
|
|
* version, so we do all of this sort of stuff here.
|
|
*/
|
|
rdi->ibdev.uverbs_cmd_mask =
|
|
(1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
|
|
(1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
|
|
(1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_AH) |
|
|
(1ull << IB_USER_VERBS_CMD_MODIFY_AH) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_AH) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_AH) |
|
|
(1ull << IB_USER_VERBS_CMD_REG_MR) |
|
|
(1ull << IB_USER_VERBS_CMD_DEREG_MR) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_RESIZE_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_POLL_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_REQ_NOTIFY_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_POST_SEND) |
|
|
(1ull << IB_USER_VERBS_CMD_POST_RECV) |
|
|
(1ull << IB_USER_VERBS_CMD_ATTACH_MCAST) |
|
|
(1ull << IB_USER_VERBS_CMD_DETACH_MCAST) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_SRQ) |
|
|
(1ull << IB_USER_VERBS_CMD_MODIFY_SRQ) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_SRQ) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_SRQ) |
|
|
(1ull << IB_USER_VERBS_CMD_POST_SRQ_RECV);
|
|
rdi->ibdev.node_type = RDMA_NODE_IB_CA;
|
|
if (!rdi->ibdev.num_comp_vectors)
|
|
rdi->ibdev.num_comp_vectors = 1;
|
|
|
|
/* We are now good to announce we exist */
|
|
ret = ib_register_device(&rdi->ibdev, dev_name(&rdi->ibdev.dev), NULL);
|
|
if (ret) {
|
|
rvt_pr_err(rdi, "Failed to register driver with ib core.\n");
|
|
goto bail_wss;
|
|
}
|
|
|
|
rvt_create_mad_agents(rdi);
|
|
|
|
rvt_pr_info(rdi, "Registration with rdmavt done.\n");
|
|
return ret;
|
|
|
|
bail_wss:
|
|
rvt_wss_exit(rdi);
|
|
bail_mr:
|
|
rvt_mr_exit(rdi);
|
|
|
|
bail_no_mr:
|
|
rvt_qp_exit(rdi);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(rvt_register_device);
|
|
|
|
/**
|
|
* rvt_unregister_device - remove a driver
|
|
* @rdi: rvt dev struct
|
|
*/
|
|
void rvt_unregister_device(struct rvt_dev_info *rdi)
|
|
{
|
|
trace_rvt_dbg(rdi, "Driver is unregistering.");
|
|
if (!rdi)
|
|
return;
|
|
|
|
rvt_free_mad_agents(rdi);
|
|
|
|
ib_unregister_device(&rdi->ibdev);
|
|
rvt_wss_exit(rdi);
|
|
rvt_mr_exit(rdi);
|
|
rvt_qp_exit(rdi);
|
|
}
|
|
EXPORT_SYMBOL(rvt_unregister_device);
|
|
|
|
/**
|
|
* rvt_init_port - init internal data for driver port
|
|
* @rdi: rvt_dev_info struct
|
|
* @port: rvt port
|
|
* @port_index: 0 based index of ports, different from IB core port num
|
|
* @pkey_table: pkey_table for @port
|
|
*
|
|
* Keep track of a list of ports. No need to have a detach port.
|
|
* They persist until the driver goes away.
|
|
*
|
|
* Return: always 0
|
|
*/
|
|
int rvt_init_port(struct rvt_dev_info *rdi, struct rvt_ibport *port,
|
|
int port_index, u16 *pkey_table)
|
|
{
|
|
|
|
rdi->ports[port_index] = port;
|
|
rdi->ports[port_index]->pkey_table = pkey_table;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(rvt_init_port);
|