In some situations a sent packet may get queued in the NIC longer than than timeout of a ULP. Currently if this happens the ULP may try to reset the link by destroying the qp and setting up an alternate connection but will fail because the rxe driver is waiting for the packet to finish getting sent and be returned to the skb destructor function where the qp reference holding things up will be dropped. This patch modifies the way that the qp is passed to the destructor to pass the qp index and not a qp pointer. Then the destructor will attempt to lookup the qp from its index and if it fails exit early. This requires taking a reference on the struct sock rather than the qp allowing the qp to be destroyed while the sk is still around waiting for the packet to finish. Link: https://lore.kernel.org/r/20240329145513.35381-15-rpearsonhpe@gmail.com Signed-off-by: Bob Pearson <rpearsonhpe@gmail.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
870 lines
20 KiB
C
870 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
|
|
/*
|
|
* Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved.
|
|
* Copyright (c) 2015 System Fabric Works, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/skbuff.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <rdma/uverbs_ioctl.h>
|
|
|
|
#include "rxe.h"
|
|
#include "rxe_loc.h"
|
|
#include "rxe_queue.h"
|
|
#include "rxe_task.h"
|
|
|
|
static int rxe_qp_chk_cap(struct rxe_dev *rxe, struct ib_qp_cap *cap,
|
|
int has_srq)
|
|
{
|
|
if (cap->max_send_wr > rxe->attr.max_qp_wr) {
|
|
rxe_dbg_dev(rxe, "invalid send wr = %u > %d\n",
|
|
cap->max_send_wr, rxe->attr.max_qp_wr);
|
|
goto err1;
|
|
}
|
|
|
|
if (cap->max_send_sge > rxe->attr.max_send_sge) {
|
|
rxe_dbg_dev(rxe, "invalid send sge = %u > %d\n",
|
|
cap->max_send_sge, rxe->attr.max_send_sge);
|
|
goto err1;
|
|
}
|
|
|
|
if (!has_srq) {
|
|
if (cap->max_recv_wr > rxe->attr.max_qp_wr) {
|
|
rxe_dbg_dev(rxe, "invalid recv wr = %u > %d\n",
|
|
cap->max_recv_wr, rxe->attr.max_qp_wr);
|
|
goto err1;
|
|
}
|
|
|
|
if (cap->max_recv_sge > rxe->attr.max_recv_sge) {
|
|
rxe_dbg_dev(rxe, "invalid recv sge = %u > %d\n",
|
|
cap->max_recv_sge, rxe->attr.max_recv_sge);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
if (cap->max_inline_data > rxe->max_inline_data) {
|
|
rxe_dbg_dev(rxe, "invalid max inline data = %u > %d\n",
|
|
cap->max_inline_data, rxe->max_inline_data);
|
|
goto err1;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err1:
|
|
return -EINVAL;
|
|
}
|
|
|
|
int rxe_qp_chk_init(struct rxe_dev *rxe, struct ib_qp_init_attr *init)
|
|
{
|
|
struct ib_qp_cap *cap = &init->cap;
|
|
struct rxe_port *port;
|
|
int port_num = init->port_num;
|
|
|
|
switch (init->qp_type) {
|
|
case IB_QPT_GSI:
|
|
case IB_QPT_RC:
|
|
case IB_QPT_UC:
|
|
case IB_QPT_UD:
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
if (!init->recv_cq || !init->send_cq) {
|
|
rxe_dbg_dev(rxe, "missing cq\n");
|
|
goto err1;
|
|
}
|
|
|
|
if (rxe_qp_chk_cap(rxe, cap, !!init->srq))
|
|
goto err1;
|
|
|
|
if (init->qp_type == IB_QPT_GSI) {
|
|
if (!rdma_is_port_valid(&rxe->ib_dev, port_num)) {
|
|
rxe_dbg_dev(rxe, "invalid port = %d\n", port_num);
|
|
goto err1;
|
|
}
|
|
|
|
port = &rxe->port;
|
|
|
|
if (init->qp_type == IB_QPT_GSI && port->qp_gsi_index) {
|
|
rxe_dbg_dev(rxe, "GSI QP exists for port %d\n", port_num);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err1:
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int alloc_rd_atomic_resources(struct rxe_qp *qp, unsigned int n)
|
|
{
|
|
qp->resp.res_head = 0;
|
|
qp->resp.res_tail = 0;
|
|
qp->resp.resources = kcalloc(n, sizeof(struct resp_res), GFP_KERNEL);
|
|
|
|
if (!qp->resp.resources)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void free_rd_atomic_resources(struct rxe_qp *qp)
|
|
{
|
|
if (qp->resp.resources) {
|
|
int i;
|
|
|
|
for (i = 0; i < qp->attr.max_dest_rd_atomic; i++) {
|
|
struct resp_res *res = &qp->resp.resources[i];
|
|
|
|
free_rd_atomic_resource(res);
|
|
}
|
|
kfree(qp->resp.resources);
|
|
qp->resp.resources = NULL;
|
|
}
|
|
}
|
|
|
|
void free_rd_atomic_resource(struct resp_res *res)
|
|
{
|
|
res->type = 0;
|
|
}
|
|
|
|
static void cleanup_rd_atomic_resources(struct rxe_qp *qp)
|
|
{
|
|
int i;
|
|
struct resp_res *res;
|
|
|
|
if (qp->resp.resources) {
|
|
for (i = 0; i < qp->attr.max_dest_rd_atomic; i++) {
|
|
res = &qp->resp.resources[i];
|
|
free_rd_atomic_resource(res);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void rxe_qp_init_misc(struct rxe_dev *rxe, struct rxe_qp *qp,
|
|
struct ib_qp_init_attr *init)
|
|
{
|
|
struct rxe_port *port;
|
|
u32 qpn;
|
|
|
|
qp->sq_sig_type = init->sq_sig_type;
|
|
qp->attr.path_mtu = 1;
|
|
qp->mtu = ib_mtu_enum_to_int(qp->attr.path_mtu);
|
|
|
|
qpn = qp->elem.index;
|
|
port = &rxe->port;
|
|
|
|
switch (init->qp_type) {
|
|
case IB_QPT_GSI:
|
|
qp->ibqp.qp_num = 1;
|
|
port->qp_gsi_index = qpn;
|
|
qp->attr.port_num = init->port_num;
|
|
break;
|
|
|
|
default:
|
|
qp->ibqp.qp_num = qpn;
|
|
break;
|
|
}
|
|
|
|
spin_lock_init(&qp->state_lock);
|
|
|
|
spin_lock_init(&qp->sq.sq_lock);
|
|
spin_lock_init(&qp->rq.producer_lock);
|
|
spin_lock_init(&qp->rq.consumer_lock);
|
|
|
|
skb_queue_head_init(&qp->req_pkts);
|
|
skb_queue_head_init(&qp->resp_pkts);
|
|
|
|
atomic_set(&qp->ssn, 0);
|
|
atomic_set(&qp->skb_out, 0);
|
|
}
|
|
|
|
static int rxe_init_sq(struct rxe_qp *qp, struct ib_qp_init_attr *init,
|
|
struct ib_udata *udata,
|
|
struct rxe_create_qp_resp __user *uresp)
|
|
{
|
|
struct rxe_dev *rxe = to_rdev(qp->ibqp.device);
|
|
int wqe_size;
|
|
int err;
|
|
|
|
qp->sq.max_wr = init->cap.max_send_wr;
|
|
wqe_size = max_t(int, init->cap.max_send_sge * sizeof(struct ib_sge),
|
|
init->cap.max_inline_data);
|
|
qp->sq.max_sge = wqe_size / sizeof(struct ib_sge);
|
|
qp->sq.max_inline = wqe_size;
|
|
wqe_size += sizeof(struct rxe_send_wqe);
|
|
|
|
qp->sq.queue = rxe_queue_init(rxe, &qp->sq.max_wr, wqe_size,
|
|
QUEUE_TYPE_FROM_CLIENT);
|
|
if (!qp->sq.queue) {
|
|
rxe_err_qp(qp, "Unable to allocate send queue\n");
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
|
|
/* prepare info for caller to mmap send queue if user space qp */
|
|
err = do_mmap_info(rxe, uresp ? &uresp->sq_mi : NULL, udata,
|
|
qp->sq.queue->buf, qp->sq.queue->buf_size,
|
|
&qp->sq.queue->ip);
|
|
if (err) {
|
|
rxe_err_qp(qp, "do_mmap_info failed, err = %d\n", err);
|
|
goto err_free;
|
|
}
|
|
|
|
/* return actual capabilities to caller which may be larger
|
|
* than requested
|
|
*/
|
|
init->cap.max_send_wr = qp->sq.max_wr;
|
|
init->cap.max_send_sge = qp->sq.max_sge;
|
|
init->cap.max_inline_data = qp->sq.max_inline;
|
|
|
|
return 0;
|
|
|
|
err_free:
|
|
vfree(qp->sq.queue->buf);
|
|
kfree(qp->sq.queue);
|
|
qp->sq.queue = NULL;
|
|
err_out:
|
|
return err;
|
|
}
|
|
|
|
static int rxe_qp_init_req(struct rxe_dev *rxe, struct rxe_qp *qp,
|
|
struct ib_qp_init_attr *init, struct ib_udata *udata,
|
|
struct rxe_create_qp_resp __user *uresp)
|
|
{
|
|
int err;
|
|
|
|
/* if we don't finish qp create make sure queue is valid */
|
|
skb_queue_head_init(&qp->req_pkts);
|
|
|
|
err = sock_create_kern(&init_net, AF_INET, SOCK_DGRAM, 0, &qp->sk);
|
|
if (err < 0)
|
|
return err;
|
|
qp->sk->sk->sk_user_data = (void *)(uintptr_t)qp->elem.index;
|
|
|
|
/* pick a source UDP port number for this QP based on
|
|
* the source QPN. this spreads traffic for different QPs
|
|
* across different NIC RX queues (while using a single
|
|
* flow for a given QP to maintain packet order).
|
|
* the port number must be in the Dynamic Ports range
|
|
* (0xc000 - 0xffff).
|
|
*/
|
|
qp->src_port = RXE_ROCE_V2_SPORT + (hash_32(qp_num(qp), 14) & 0x3fff);
|
|
|
|
err = rxe_init_sq(qp, init, udata, uresp);
|
|
if (err)
|
|
return err;
|
|
|
|
qp->req.wqe_index = queue_get_producer(qp->sq.queue,
|
|
QUEUE_TYPE_FROM_CLIENT);
|
|
|
|
qp->req.opcode = -1;
|
|
qp->comp.opcode = -1;
|
|
|
|
rxe_init_task(&qp->send_task, qp, rxe_sender);
|
|
|
|
qp->qp_timeout_jiffies = 0; /* Can't be set for UD/UC in modify_qp */
|
|
if (init->qp_type == IB_QPT_RC) {
|
|
timer_setup(&qp->rnr_nak_timer, rnr_nak_timer, 0);
|
|
timer_setup(&qp->retrans_timer, retransmit_timer, 0);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int rxe_init_rq(struct rxe_qp *qp, struct ib_qp_init_attr *init,
|
|
struct ib_udata *udata,
|
|
struct rxe_create_qp_resp __user *uresp)
|
|
{
|
|
struct rxe_dev *rxe = to_rdev(qp->ibqp.device);
|
|
int wqe_size;
|
|
int err;
|
|
|
|
qp->rq.max_wr = init->cap.max_recv_wr;
|
|
qp->rq.max_sge = init->cap.max_recv_sge;
|
|
wqe_size = sizeof(struct rxe_recv_wqe) +
|
|
qp->rq.max_sge*sizeof(struct ib_sge);
|
|
|
|
qp->rq.queue = rxe_queue_init(rxe, &qp->rq.max_wr, wqe_size,
|
|
QUEUE_TYPE_FROM_CLIENT);
|
|
if (!qp->rq.queue) {
|
|
rxe_err_qp(qp, "Unable to allocate recv queue\n");
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
|
|
/* prepare info for caller to mmap recv queue if user space qp */
|
|
err = do_mmap_info(rxe, uresp ? &uresp->rq_mi : NULL, udata,
|
|
qp->rq.queue->buf, qp->rq.queue->buf_size,
|
|
&qp->rq.queue->ip);
|
|
if (err) {
|
|
rxe_err_qp(qp, "do_mmap_info failed, err = %d\n", err);
|
|
goto err_free;
|
|
}
|
|
|
|
/* return actual capabilities to caller which may be larger
|
|
* than requested
|
|
*/
|
|
init->cap.max_recv_wr = qp->rq.max_wr;
|
|
|
|
return 0;
|
|
|
|
err_free:
|
|
vfree(qp->rq.queue->buf);
|
|
kfree(qp->rq.queue);
|
|
qp->rq.queue = NULL;
|
|
err_out:
|
|
return err;
|
|
}
|
|
|
|
static int rxe_qp_init_resp(struct rxe_dev *rxe, struct rxe_qp *qp,
|
|
struct ib_qp_init_attr *init,
|
|
struct ib_udata *udata,
|
|
struct rxe_create_qp_resp __user *uresp)
|
|
{
|
|
int err;
|
|
|
|
/* if we don't finish qp create make sure queue is valid */
|
|
skb_queue_head_init(&qp->resp_pkts);
|
|
|
|
if (!qp->srq) {
|
|
err = rxe_init_rq(qp, init, udata, uresp);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
rxe_init_task(&qp->recv_task, qp, rxe_receiver);
|
|
|
|
qp->resp.opcode = OPCODE_NONE;
|
|
qp->resp.msn = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* called by the create qp verb */
|
|
int rxe_qp_from_init(struct rxe_dev *rxe, struct rxe_qp *qp, struct rxe_pd *pd,
|
|
struct ib_qp_init_attr *init,
|
|
struct rxe_create_qp_resp __user *uresp,
|
|
struct ib_pd *ibpd,
|
|
struct ib_udata *udata)
|
|
{
|
|
int err;
|
|
struct rxe_cq *rcq = to_rcq(init->recv_cq);
|
|
struct rxe_cq *scq = to_rcq(init->send_cq);
|
|
struct rxe_srq *srq = init->srq ? to_rsrq(init->srq) : NULL;
|
|
unsigned long flags;
|
|
|
|
rxe_get(pd);
|
|
rxe_get(rcq);
|
|
rxe_get(scq);
|
|
if (srq)
|
|
rxe_get(srq);
|
|
|
|
qp->pd = pd;
|
|
qp->rcq = rcq;
|
|
qp->scq = scq;
|
|
qp->srq = srq;
|
|
|
|
atomic_inc(&rcq->num_wq);
|
|
atomic_inc(&scq->num_wq);
|
|
|
|
rxe_qp_init_misc(rxe, qp, init);
|
|
|
|
err = rxe_qp_init_req(rxe, qp, init, udata, uresp);
|
|
if (err)
|
|
goto err1;
|
|
|
|
err = rxe_qp_init_resp(rxe, qp, init, udata, uresp);
|
|
if (err)
|
|
goto err2;
|
|
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
qp->attr.qp_state = IB_QPS_RESET;
|
|
qp->valid = 1;
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
|
|
return 0;
|
|
|
|
err2:
|
|
rxe_queue_cleanup(qp->sq.queue);
|
|
qp->sq.queue = NULL;
|
|
err1:
|
|
atomic_dec(&rcq->num_wq);
|
|
atomic_dec(&scq->num_wq);
|
|
|
|
qp->pd = NULL;
|
|
qp->rcq = NULL;
|
|
qp->scq = NULL;
|
|
qp->srq = NULL;
|
|
|
|
if (srq)
|
|
rxe_put(srq);
|
|
rxe_put(scq);
|
|
rxe_put(rcq);
|
|
rxe_put(pd);
|
|
|
|
return err;
|
|
}
|
|
|
|
/* called by the query qp verb */
|
|
int rxe_qp_to_init(struct rxe_qp *qp, struct ib_qp_init_attr *init)
|
|
{
|
|
init->event_handler = qp->ibqp.event_handler;
|
|
init->qp_context = qp->ibqp.qp_context;
|
|
init->send_cq = qp->ibqp.send_cq;
|
|
init->recv_cq = qp->ibqp.recv_cq;
|
|
init->srq = qp->ibqp.srq;
|
|
|
|
init->cap.max_send_wr = qp->sq.max_wr;
|
|
init->cap.max_send_sge = qp->sq.max_sge;
|
|
init->cap.max_inline_data = qp->sq.max_inline;
|
|
|
|
if (!qp->srq) {
|
|
init->cap.max_recv_wr = qp->rq.max_wr;
|
|
init->cap.max_recv_sge = qp->rq.max_sge;
|
|
}
|
|
|
|
init->sq_sig_type = qp->sq_sig_type;
|
|
|
|
init->qp_type = qp->ibqp.qp_type;
|
|
init->port_num = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int rxe_qp_chk_attr(struct rxe_dev *rxe, struct rxe_qp *qp,
|
|
struct ib_qp_attr *attr, int mask)
|
|
{
|
|
if (mask & IB_QP_PORT) {
|
|
if (!rdma_is_port_valid(&rxe->ib_dev, attr->port_num)) {
|
|
rxe_dbg_qp(qp, "invalid port %d\n", attr->port_num);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_CAP && rxe_qp_chk_cap(rxe, &attr->cap, !!qp->srq))
|
|
goto err1;
|
|
|
|
if (mask & IB_QP_ACCESS_FLAGS) {
|
|
if (!(qp_type(qp) == IB_QPT_RC || qp_type(qp) == IB_QPT_UC))
|
|
goto err1;
|
|
if (attr->qp_access_flags & ~RXE_ACCESS_SUPPORTED_QP)
|
|
goto err1;
|
|
}
|
|
|
|
if (mask & IB_QP_AV && rxe_av_chk_attr(qp, &attr->ah_attr))
|
|
goto err1;
|
|
|
|
if (mask & IB_QP_ALT_PATH) {
|
|
if (rxe_av_chk_attr(qp, &attr->alt_ah_attr))
|
|
goto err1;
|
|
if (!rdma_is_port_valid(&rxe->ib_dev, attr->alt_port_num)) {
|
|
rxe_dbg_qp(qp, "invalid alt port %d\n", attr->alt_port_num);
|
|
goto err1;
|
|
}
|
|
if (attr->alt_timeout > 31) {
|
|
rxe_dbg_qp(qp, "invalid alt timeout %d > 31\n",
|
|
attr->alt_timeout);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_PATH_MTU) {
|
|
struct rxe_port *port = &rxe->port;
|
|
|
|
enum ib_mtu max_mtu = port->attr.max_mtu;
|
|
enum ib_mtu mtu = attr->path_mtu;
|
|
|
|
if (mtu > max_mtu) {
|
|
rxe_dbg_qp(qp, "invalid mtu (%d) > (%d)\n",
|
|
ib_mtu_enum_to_int(mtu),
|
|
ib_mtu_enum_to_int(max_mtu));
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_MAX_QP_RD_ATOMIC) {
|
|
if (attr->max_rd_atomic > rxe->attr.max_qp_rd_atom) {
|
|
rxe_dbg_qp(qp, "invalid max_rd_atomic %d > %d\n",
|
|
attr->max_rd_atomic,
|
|
rxe->attr.max_qp_rd_atom);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_TIMEOUT) {
|
|
if (attr->timeout > 31) {
|
|
rxe_dbg_qp(qp, "invalid timeout %d > 31\n",
|
|
attr->timeout);
|
|
goto err1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err1:
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* move the qp to the reset state */
|
|
static void rxe_qp_reset(struct rxe_qp *qp)
|
|
{
|
|
/* stop tasks from running */
|
|
rxe_disable_task(&qp->recv_task);
|
|
rxe_disable_task(&qp->send_task);
|
|
|
|
/* drain work and packet queuesc */
|
|
rxe_sender(qp);
|
|
rxe_receiver(qp);
|
|
|
|
if (qp->rq.queue)
|
|
rxe_queue_reset(qp->rq.queue);
|
|
if (qp->sq.queue)
|
|
rxe_queue_reset(qp->sq.queue);
|
|
|
|
/* cleanup attributes */
|
|
atomic_set(&qp->ssn, 0);
|
|
qp->req.opcode = -1;
|
|
qp->req.need_retry = 0;
|
|
qp->req.wait_for_rnr_timer = 0;
|
|
qp->req.noack_pkts = 0;
|
|
qp->resp.msn = 0;
|
|
qp->resp.opcode = -1;
|
|
qp->resp.drop_msg = 0;
|
|
qp->resp.goto_error = 0;
|
|
qp->resp.sent_psn_nak = 0;
|
|
|
|
if (qp->resp.mr) {
|
|
rxe_put(qp->resp.mr);
|
|
qp->resp.mr = NULL;
|
|
}
|
|
|
|
cleanup_rd_atomic_resources(qp);
|
|
|
|
/* reenable tasks */
|
|
rxe_enable_task(&qp->recv_task);
|
|
rxe_enable_task(&qp->send_task);
|
|
}
|
|
|
|
/* move the qp to the error state */
|
|
void rxe_qp_error(struct rxe_qp *qp)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
qp->attr.qp_state = IB_QPS_ERR;
|
|
|
|
/* drain work and packet queues */
|
|
rxe_sched_task(&qp->recv_task);
|
|
rxe_sched_task(&qp->send_task);
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
}
|
|
|
|
static void rxe_qp_sqd(struct rxe_qp *qp, struct ib_qp_attr *attr,
|
|
int mask)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
qp->attr.sq_draining = 1;
|
|
rxe_sched_task(&qp->send_task);
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
}
|
|
|
|
/* caller should hold qp->state_lock */
|
|
static int __qp_chk_state(struct rxe_qp *qp, struct ib_qp_attr *attr,
|
|
int mask)
|
|
{
|
|
enum ib_qp_state cur_state;
|
|
enum ib_qp_state new_state;
|
|
|
|
cur_state = (mask & IB_QP_CUR_STATE) ?
|
|
attr->cur_qp_state : qp->attr.qp_state;
|
|
new_state = (mask & IB_QP_STATE) ?
|
|
attr->qp_state : cur_state;
|
|
|
|
if (!ib_modify_qp_is_ok(cur_state, new_state, qp_type(qp), mask))
|
|
return -EINVAL;
|
|
|
|
if (mask & IB_QP_STATE && cur_state == IB_QPS_SQD) {
|
|
if (qp->attr.sq_draining && new_state != IB_QPS_ERR)
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const char *const qps2str[] = {
|
|
[IB_QPS_RESET] = "RESET",
|
|
[IB_QPS_INIT] = "INIT",
|
|
[IB_QPS_RTR] = "RTR",
|
|
[IB_QPS_RTS] = "RTS",
|
|
[IB_QPS_SQD] = "SQD",
|
|
[IB_QPS_SQE] = "SQE",
|
|
[IB_QPS_ERR] = "ERR",
|
|
};
|
|
|
|
/* called by the modify qp verb */
|
|
int rxe_qp_from_attr(struct rxe_qp *qp, struct ib_qp_attr *attr, int mask,
|
|
struct ib_udata *udata)
|
|
{
|
|
int err;
|
|
|
|
if (mask & IB_QP_CUR_STATE)
|
|
qp->attr.cur_qp_state = attr->qp_state;
|
|
|
|
if (mask & IB_QP_STATE) {
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
err = __qp_chk_state(qp, attr, mask);
|
|
if (!err) {
|
|
qp->attr.qp_state = attr->qp_state;
|
|
rxe_dbg_qp(qp, "state -> %s\n",
|
|
qps2str[attr->qp_state]);
|
|
}
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
|
|
if (err)
|
|
return err;
|
|
|
|
switch (attr->qp_state) {
|
|
case IB_QPS_RESET:
|
|
rxe_qp_reset(qp);
|
|
break;
|
|
case IB_QPS_SQD:
|
|
rxe_qp_sqd(qp, attr, mask);
|
|
break;
|
|
case IB_QPS_ERR:
|
|
rxe_qp_error(qp);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_MAX_QP_RD_ATOMIC) {
|
|
int max_rd_atomic = attr->max_rd_atomic ?
|
|
roundup_pow_of_two(attr->max_rd_atomic) : 0;
|
|
|
|
qp->attr.max_rd_atomic = max_rd_atomic;
|
|
atomic_set(&qp->req.rd_atomic, max_rd_atomic);
|
|
}
|
|
|
|
if (mask & IB_QP_MAX_DEST_RD_ATOMIC) {
|
|
int max_dest_rd_atomic = attr->max_dest_rd_atomic ?
|
|
roundup_pow_of_two(attr->max_dest_rd_atomic) : 0;
|
|
|
|
qp->attr.max_dest_rd_atomic = max_dest_rd_atomic;
|
|
|
|
free_rd_atomic_resources(qp);
|
|
|
|
err = alloc_rd_atomic_resources(qp, max_dest_rd_atomic);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
if (mask & IB_QP_EN_SQD_ASYNC_NOTIFY)
|
|
qp->attr.en_sqd_async_notify = attr->en_sqd_async_notify;
|
|
|
|
if (mask & IB_QP_ACCESS_FLAGS)
|
|
qp->attr.qp_access_flags = attr->qp_access_flags;
|
|
|
|
if (mask & IB_QP_PKEY_INDEX)
|
|
qp->attr.pkey_index = attr->pkey_index;
|
|
|
|
if (mask & IB_QP_PORT)
|
|
qp->attr.port_num = attr->port_num;
|
|
|
|
if (mask & IB_QP_QKEY)
|
|
qp->attr.qkey = attr->qkey;
|
|
|
|
if (mask & IB_QP_AV)
|
|
rxe_init_av(&attr->ah_attr, &qp->pri_av);
|
|
|
|
if (mask & IB_QP_ALT_PATH) {
|
|
rxe_init_av(&attr->alt_ah_attr, &qp->alt_av);
|
|
qp->attr.alt_port_num = attr->alt_port_num;
|
|
qp->attr.alt_pkey_index = attr->alt_pkey_index;
|
|
qp->attr.alt_timeout = attr->alt_timeout;
|
|
}
|
|
|
|
if (mask & IB_QP_PATH_MTU) {
|
|
qp->attr.path_mtu = attr->path_mtu;
|
|
qp->mtu = ib_mtu_enum_to_int(attr->path_mtu);
|
|
}
|
|
|
|
if (mask & IB_QP_TIMEOUT) {
|
|
qp->attr.timeout = attr->timeout;
|
|
if (attr->timeout == 0) {
|
|
qp->qp_timeout_jiffies = 0;
|
|
} else {
|
|
/* According to the spec, timeout = 4.096 * 2 ^ attr->timeout [us] */
|
|
int j = nsecs_to_jiffies(4096ULL << attr->timeout);
|
|
|
|
qp->qp_timeout_jiffies = j ? j : 1;
|
|
}
|
|
}
|
|
|
|
if (mask & IB_QP_RETRY_CNT) {
|
|
qp->attr.retry_cnt = attr->retry_cnt;
|
|
qp->comp.retry_cnt = attr->retry_cnt;
|
|
rxe_dbg_qp(qp, "set retry count = %d\n", attr->retry_cnt);
|
|
}
|
|
|
|
if (mask & IB_QP_RNR_RETRY) {
|
|
qp->attr.rnr_retry = attr->rnr_retry;
|
|
qp->comp.rnr_retry = attr->rnr_retry;
|
|
rxe_dbg_qp(qp, "set rnr retry count = %d\n", attr->rnr_retry);
|
|
}
|
|
|
|
if (mask & IB_QP_RQ_PSN) {
|
|
qp->attr.rq_psn = (attr->rq_psn & BTH_PSN_MASK);
|
|
qp->resp.psn = qp->attr.rq_psn;
|
|
rxe_dbg_qp(qp, "set resp psn = 0x%x\n", qp->resp.psn);
|
|
}
|
|
|
|
if (mask & IB_QP_MIN_RNR_TIMER) {
|
|
qp->attr.min_rnr_timer = attr->min_rnr_timer;
|
|
rxe_dbg_qp(qp, "set min rnr timer = 0x%x\n",
|
|
attr->min_rnr_timer);
|
|
}
|
|
|
|
if (mask & IB_QP_SQ_PSN) {
|
|
qp->attr.sq_psn = (attr->sq_psn & BTH_PSN_MASK);
|
|
qp->req.psn = qp->attr.sq_psn;
|
|
qp->comp.psn = qp->attr.sq_psn;
|
|
rxe_dbg_qp(qp, "set req psn = 0x%x\n", qp->req.psn);
|
|
}
|
|
|
|
if (mask & IB_QP_PATH_MIG_STATE)
|
|
qp->attr.path_mig_state = attr->path_mig_state;
|
|
|
|
if (mask & IB_QP_DEST_QPN)
|
|
qp->attr.dest_qp_num = attr->dest_qp_num;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* called by the query qp verb */
|
|
int rxe_qp_to_attr(struct rxe_qp *qp, struct ib_qp_attr *attr, int mask)
|
|
{
|
|
unsigned long flags;
|
|
|
|
*attr = qp->attr;
|
|
|
|
attr->rq_psn = qp->resp.psn;
|
|
attr->sq_psn = qp->req.psn;
|
|
|
|
attr->cap.max_send_wr = qp->sq.max_wr;
|
|
attr->cap.max_send_sge = qp->sq.max_sge;
|
|
attr->cap.max_inline_data = qp->sq.max_inline;
|
|
|
|
if (!qp->srq) {
|
|
attr->cap.max_recv_wr = qp->rq.max_wr;
|
|
attr->cap.max_recv_sge = qp->rq.max_sge;
|
|
}
|
|
|
|
rxe_av_to_attr(&qp->pri_av, &attr->ah_attr);
|
|
rxe_av_to_attr(&qp->alt_av, &attr->alt_ah_attr);
|
|
|
|
/* Applications that get this state typically spin on it.
|
|
* Yield the processor
|
|
*/
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
if (qp->attr.sq_draining) {
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
cond_resched();
|
|
} else {
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int rxe_qp_chk_destroy(struct rxe_qp *qp)
|
|
{
|
|
/* See IBA o10-2.2.3
|
|
* An attempt to destroy a QP while attached to a mcast group
|
|
* will fail immediately.
|
|
*/
|
|
if (atomic_read(&qp->mcg_num)) {
|
|
rxe_dbg_qp(qp, "Attempt to destroy while attached to multicast group\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* called when the last reference to the qp is dropped */
|
|
static void rxe_qp_do_cleanup(struct work_struct *work)
|
|
{
|
|
struct rxe_qp *qp = container_of(work, typeof(*qp), cleanup_work.work);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&qp->state_lock, flags);
|
|
qp->valid = 0;
|
|
spin_unlock_irqrestore(&qp->state_lock, flags);
|
|
qp->qp_timeout_jiffies = 0;
|
|
|
|
if (qp_type(qp) == IB_QPT_RC) {
|
|
del_timer_sync(&qp->retrans_timer);
|
|
del_timer_sync(&qp->rnr_nak_timer);
|
|
}
|
|
|
|
if (qp->recv_task.func)
|
|
rxe_cleanup_task(&qp->recv_task);
|
|
|
|
if (qp->send_task.func)
|
|
rxe_cleanup_task(&qp->send_task);
|
|
|
|
/* flush out any receive wr's or pending requests */
|
|
rxe_sender(qp);
|
|
rxe_receiver(qp);
|
|
|
|
if (qp->sq.queue)
|
|
rxe_queue_cleanup(qp->sq.queue);
|
|
|
|
if (qp->srq)
|
|
rxe_put(qp->srq);
|
|
|
|
if (qp->rq.queue)
|
|
rxe_queue_cleanup(qp->rq.queue);
|
|
|
|
if (qp->scq) {
|
|
atomic_dec(&qp->scq->num_wq);
|
|
rxe_put(qp->scq);
|
|
}
|
|
|
|
if (qp->rcq) {
|
|
atomic_dec(&qp->rcq->num_wq);
|
|
rxe_put(qp->rcq);
|
|
}
|
|
|
|
if (qp->pd)
|
|
rxe_put(qp->pd);
|
|
|
|
if (qp->resp.mr)
|
|
rxe_put(qp->resp.mr);
|
|
|
|
free_rd_atomic_resources(qp);
|
|
|
|
if (qp->sk) {
|
|
if (qp_type(qp) == IB_QPT_RC)
|
|
sk_dst_reset(qp->sk->sk);
|
|
|
|
kernel_sock_shutdown(qp->sk, SHUT_RDWR);
|
|
sock_release(qp->sk);
|
|
}
|
|
}
|
|
|
|
/* called when the last reference to the qp is dropped */
|
|
void rxe_qp_cleanup(struct rxe_pool_elem *elem)
|
|
{
|
|
struct rxe_qp *qp = container_of(elem, typeof(*qp), elem);
|
|
|
|
execute_in_process_context(rxe_qp_do_cleanup, &qp->cleanup_work);
|
|
}
|