freebsd-src/contrib/ofed/libcxgb4/verbs.c
Sean Lim a687910fc4 Cleanup pthread locks in ofed RDMA verbs
On FreeBSD, pthread mutex, cond, and spinlocks allocate memory.  On
Linux-based systems, these calls do not allocate memory.  So there was a
safe assumption that the ofed RDMA verb calls do not need to explicitly
destroy the pthread locks.  This assumption is false on FreeBSD.  So let
us rearrange the code to cleanup the pthread locks.

Reviewed by:	delphij
MFC after:	2 weeks
Sponsored by:	Dell EMC Isilon
Differential Revision:	https://reviews.freebsd.org/D41105
2023-09-19 09:10:42 -05:00

726 lines
18 KiB
C

/*
* Copyright (c) 2006-2016 Chelsio, Inc. All rights reserved.
*
* This software is available to you under a choice of one of two
* licenses. You may choose to be licensed under the terms of the GNU
* General Public License (GPL) Version 2, available from the file
* COPYING in the main directory of this source tree, or the
* OpenIB.org BSD license below:
*
* Redistribution and use in source and binary forms, with or
* without modification, are permitted provided that the following
* conditions are met:
*
* - Redistributions of source code must retain the above
* copyright notice, this list of conditions and the following
* disclaimer.
*
* - Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials
* provided with the distribution.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
#include <config.h>
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <errno.h>
#include <pthread.h>
#include <sys/mman.h>
#include <inttypes.h>
#include <assert.h>
#include "libcxgb4.h"
#include "cxgb4-abi.h"
#define MASKED(x) (void *)((unsigned long)(x) & c4iw_page_mask)
int c4iw_query_device(struct ibv_context *context, struct ibv_device_attr *attr)
{
struct ibv_query_device cmd;
uint64_t raw_fw_ver;
u8 major, minor, sub_minor, build;
int ret;
ret = ibv_cmd_query_device(context, attr, &raw_fw_ver, &cmd,
sizeof cmd);
if (ret)
return ret;
major = (raw_fw_ver >> 24) & 0xff;
minor = (raw_fw_ver >> 16) & 0xff;
sub_minor = (raw_fw_ver >> 8) & 0xff;
build = raw_fw_ver & 0xff;
snprintf(attr->fw_ver, sizeof attr->fw_ver,
"%d.%d.%d.%d", major, minor, sub_minor, build);
return 0;
}
int c4iw_query_port(struct ibv_context *context, uint8_t port,
struct ibv_port_attr *attr)
{
struct ibv_query_port cmd;
return ibv_cmd_query_port(context, port, attr, &cmd, sizeof cmd);
}
struct ibv_pd *c4iw_alloc_pd(struct ibv_context *context)
{
struct ibv_alloc_pd cmd;
struct c4iw_alloc_pd_resp resp;
struct c4iw_pd *pd;
pd = malloc(sizeof *pd);
if (!pd)
return NULL;
if (ibv_cmd_alloc_pd(context, &pd->ibv_pd, &cmd, sizeof cmd,
&resp.ibv_resp, sizeof resp)) {
free(pd);
return NULL;
}
return &pd->ibv_pd;
}
int c4iw_free_pd(struct ibv_pd *pd)
{
int ret;
ret = ibv_cmd_dealloc_pd(pd);
if (ret)
return ret;
free(pd);
return 0;
}
static struct ibv_mr *__c4iw_reg_mr(struct ibv_pd *pd, void *addr,
size_t length, uint64_t hca_va,
int access)
{
struct c4iw_mr *mhp;
struct ibv_reg_mr cmd;
struct ibv_reg_mr_resp resp;
struct c4iw_dev *dev = to_c4iw_dev(pd->context->device);
mhp = malloc(sizeof *mhp);
if (!mhp)
return NULL;
if (ibv_cmd_reg_mr(pd, addr, length, hca_va,
access, &mhp->ibv_mr, &cmd, sizeof cmd,
&resp, sizeof resp)) {
free(mhp);
return NULL;
}
mhp->va_fbo = hca_va;
mhp->len = length;
PDBG("%s stag 0x%x va_fbo 0x%" PRIx64 " len %d\n",
__func__, mhp->ibv_mr.rkey, mhp->va_fbo, mhp->len);
pthread_spin_lock(&dev->lock);
dev->mmid2ptr[c4iw_mmid(mhp->ibv_mr.lkey)] = mhp;
pthread_spin_unlock(&dev->lock);
INC_STAT(mr);
return &mhp->ibv_mr;
}
struct ibv_mr *c4iw_reg_mr(struct ibv_pd *pd, void *addr,
size_t length, int access)
{
PDBG("%s addr %p length %ld\n", __func__, addr, length);
return __c4iw_reg_mr(pd, addr, length, (uintptr_t) addr, access);
}
int c4iw_dereg_mr(struct ibv_mr *mr)
{
int ret;
struct c4iw_dev *dev = to_c4iw_dev(mr->pd->context->device);
ret = ibv_cmd_dereg_mr(mr);
if (ret)
return ret;
pthread_spin_lock(&dev->lock);
dev->mmid2ptr[c4iw_mmid(mr->lkey)] = NULL;
pthread_spin_unlock(&dev->lock);
free(to_c4iw_mr(mr));
return 0;
}
struct ibv_cq *c4iw_create_cq(struct ibv_context *context, int cqe,
struct ibv_comp_channel *channel, int comp_vector)
{
struct ibv_create_cq cmd;
struct c4iw_create_cq_resp resp;
struct c4iw_cq *chp;
struct c4iw_dev *dev = to_c4iw_dev(context->device);
int ret;
chp = calloc(1, sizeof *chp);
if (!chp) {
return NULL;
}
resp.reserved = 0;
ret = ibv_cmd_create_cq(context, cqe, channel, comp_vector,
&chp->ibv_cq, &cmd, sizeof cmd,
&resp.ibv_resp, sizeof resp);
if (ret)
goto err1;
if (resp.reserved)
PDBG("%s c4iw_create_cq_resp reserved field modified by kernel\n",
__FUNCTION__);
ret = pthread_spin_init(&chp->lock, PTHREAD_PROCESS_PRIVATE);
if (ret)
goto err2;
#ifdef STALL_DETECTION
gettimeofday(&chp->time, NULL);
#endif
chp->rhp = dev;
chp->cq.qid_mask = resp.qid_mask;
chp->cq.cqid = resp.cqid;
chp->cq.size = resp.size;
chp->cq.memsize = resp.memsize;
chp->cq.gen = 1;
chp->cq.queue = mmap(NULL, chp->cq.memsize, PROT_READ|PROT_WRITE,
MAP_SHARED, context->cmd_fd, resp.key);
if (chp->cq.queue == MAP_FAILED)
goto err3;
chp->cq.ugts = mmap(NULL, c4iw_page_size, PROT_WRITE, MAP_SHARED,
context->cmd_fd, resp.gts_key);
if (chp->cq.ugts == MAP_FAILED)
goto err4;
if (dev_is_t4(chp->rhp))
chp->cq.ugts += 1;
else
chp->cq.ugts += 5;
chp->cq.sw_queue = calloc(chp->cq.size, sizeof *chp->cq.queue);
if (!chp->cq.sw_queue)
goto err5;
PDBG("%s cqid 0x%x key %" PRIx64 " va %p memsize %lu gts_key %"
PRIx64 " va %p qid_mask 0x%x\n",
__func__, chp->cq.cqid, resp.key, chp->cq.queue,
chp->cq.memsize, resp.gts_key, chp->cq.ugts, chp->cq.qid_mask);
pthread_spin_lock(&dev->lock);
dev->cqid2ptr[chp->cq.cqid] = chp;
pthread_spin_unlock(&dev->lock);
INC_STAT(cq);
return &chp->ibv_cq;
err5:
munmap(MASKED(chp->cq.ugts), c4iw_page_size);
err4:
munmap(chp->cq.queue, chp->cq.memsize);
err3:
pthread_spin_destroy(&chp->lock);
err2:
(void)ibv_cmd_destroy_cq(&chp->ibv_cq);
err1:
free(chp);
return NULL;
}
int c4iw_resize_cq(struct ibv_cq *ibcq, int cqe)
{
#if 0
int ret;
struct ibv_resize_cq cmd;
struct ibv_resize_cq_resp resp;
ret = ibv_cmd_resize_cq(ibcq, cqe, &cmd, sizeof cmd, &resp, sizeof resp);
PDBG("%s ret %d\n", __func__, ret);
return ret;
#else
return -ENOSYS;
#endif
}
int c4iw_destroy_cq(struct ibv_cq *ibcq)
{
int ret;
struct c4iw_cq *chp = to_c4iw_cq(ibcq);
struct c4iw_dev *dev = to_c4iw_dev(ibcq->context->device);
chp->cq.error = 1;
ret = ibv_cmd_destroy_cq(ibcq);
if (ret) {
return ret;
}
verbs_cleanup_cq(ibcq);
munmap(MASKED(chp->cq.ugts), c4iw_page_size);
munmap(chp->cq.queue, chp->cq.memsize);
pthread_spin_lock(&dev->lock);
dev->cqid2ptr[chp->cq.cqid] = NULL;
pthread_spin_unlock(&dev->lock);
free(chp->cq.sw_queue);
pthread_spin_destroy(&chp->lock);
free(chp);
return 0;
}
struct ibv_srq *c4iw_create_srq(struct ibv_pd *pd,
struct ibv_srq_init_attr *attr)
{
return NULL;
}
int c4iw_modify_srq(struct ibv_srq *srq, struct ibv_srq_attr *attr,
int attr_mask)
{
return ENOSYS;
}
int c4iw_destroy_srq(struct ibv_srq *srq)
{
return ENOSYS;
}
int c4iw_post_srq_recv(struct ibv_srq *ibsrq, struct ibv_recv_wr *wr,
struct ibv_recv_wr **bad_wr)
{
return ENOSYS;
}
static struct ibv_qp *create_qp_v0(struct ibv_pd *pd,
struct ibv_qp_init_attr *attr)
{
struct ibv_create_qp cmd;
struct c4iw_create_qp_resp_v0 resp;
struct c4iw_qp *qhp;
struct c4iw_dev *dev = to_c4iw_dev(pd->context->device);
int ret;
void *dbva;
PDBG("%s enter qp\n", __func__);
qhp = calloc(1, sizeof *qhp);
if (!qhp)
goto err1;
ret = ibv_cmd_create_qp(pd, &qhp->ibv_qp, attr, &cmd,
sizeof cmd, &resp.ibv_resp, sizeof resp);
if (ret)
goto err2;
PDBG("%s sqid 0x%x sq key %" PRIx64 " sq db/gts key %" PRIx64
" rqid 0x%x rq key %" PRIx64 " rq db/gts key %" PRIx64
" qid_mask 0x%x\n",
__func__,
resp.sqid, resp.sq_key, resp.sq_db_gts_key,
resp.rqid, resp.rq_key, resp.rq_db_gts_key, resp.qid_mask);
qhp->wq.qid_mask = resp.qid_mask;
qhp->rhp = dev;
qhp->wq.sq.qid = resp.sqid;
qhp->wq.sq.size = resp.sq_size;
qhp->wq.sq.memsize = resp.sq_memsize;
qhp->wq.sq.flags = 0;
qhp->wq.rq.msn = 1;
qhp->wq.rq.qid = resp.rqid;
qhp->wq.rq.size = resp.rq_size;
qhp->wq.rq.memsize = resp.rq_memsize;
ret = pthread_spin_init(&qhp->lock, PTHREAD_PROCESS_PRIVATE);
if (ret)
goto err3;
dbva = mmap(NULL, c4iw_page_size, PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.sq_db_gts_key);
if (dbva == MAP_FAILED)
goto err4;
qhp->wq.sq.udb = dbva;
qhp->wq.sq.queue = mmap(NULL, qhp->wq.sq.memsize,
PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.sq_key);
if (qhp->wq.sq.queue == MAP_FAILED)
goto err5;
dbva = mmap(NULL, c4iw_page_size, PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.rq_db_gts_key);
if (dbva == MAP_FAILED)
goto err6;
qhp->wq.rq.udb = dbva;
qhp->wq.rq.queue = mmap(NULL, qhp->wq.rq.memsize,
PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.rq_key);
if (qhp->wq.rq.queue == MAP_FAILED)
goto err7;
qhp->wq.sq.sw_sq = calloc(qhp->wq.sq.size, sizeof (struct t4_swsqe));
if (!qhp->wq.sq.sw_sq)
goto err8;
qhp->wq.rq.sw_rq = calloc(qhp->wq.rq.size, sizeof (uint64_t));
if (!qhp->wq.rq.sw_rq)
goto err9;
PDBG("%s sq dbva %p sq qva %p sq depth %u sq memsize %lu "
" rq dbva %p rq qva %p rq depth %u rq memsize %lu\n",
__func__,
qhp->wq.sq.udb, qhp->wq.sq.queue,
qhp->wq.sq.size, qhp->wq.sq.memsize,
qhp->wq.rq.udb, qhp->wq.rq.queue,
qhp->wq.rq.size, qhp->wq.rq.memsize);
qhp->sq_sig_all = attr->sq_sig_all;
pthread_spin_lock(&dev->lock);
dev->qpid2ptr[qhp->wq.sq.qid] = qhp;
pthread_spin_unlock(&dev->lock);
INC_STAT(qp);
return &qhp->ibv_qp;
err9:
free(qhp->wq.sq.sw_sq);
err8:
munmap((void *)qhp->wq.rq.queue, qhp->wq.rq.memsize);
err7:
munmap(MASKED(qhp->wq.rq.udb), c4iw_page_size);
err6:
munmap((void *)qhp->wq.sq.queue, qhp->wq.sq.memsize);
err5:
munmap(MASKED(qhp->wq.sq.udb), c4iw_page_size);
err4:
pthread_spin_destroy(&qhp->lock);
err3:
(void)ibv_cmd_destroy_qp(&qhp->ibv_qp);
err2:
free(qhp);
err1:
return NULL;
}
static struct ibv_qp *create_qp(struct ibv_pd *pd,
struct ibv_qp_init_attr *attr)
{
struct ibv_create_qp cmd;
struct c4iw_create_qp_resp resp;
struct c4iw_qp *qhp;
struct c4iw_dev *dev = to_c4iw_dev(pd->context->device);
struct c4iw_context *ctx = to_c4iw_context(pd->context);
int ret;
void *dbva;
PDBG("%s enter qp\n", __func__);
qhp = calloc(1, sizeof *qhp);
if (!qhp)
goto err1;
ret = ibv_cmd_create_qp(pd, &qhp->ibv_qp, attr, &cmd,
sizeof cmd, &resp.ibv_resp, sizeof resp);
if (ret)
goto err2;
PDBG("%s sqid 0x%x sq key %" PRIx64 " sq db/gts key %" PRIx64
" rqid 0x%x rq key %" PRIx64 " rq db/gts key %" PRIx64
" qid_mask 0x%x\n",
__func__,
resp.sqid, resp.sq_key, resp.sq_db_gts_key,
resp.rqid, resp.rq_key, resp.rq_db_gts_key, resp.qid_mask);
qhp->wq.qid_mask = resp.qid_mask;
qhp->rhp = dev;
qhp->wq.sq.qid = resp.sqid;
qhp->wq.sq.size = resp.sq_size;
qhp->wq.sq.memsize = resp.sq_memsize;
qhp->wq.sq.flags = resp.flags & C4IW_QPF_ONCHIP ? T4_SQ_ONCHIP : 0;
qhp->wq.sq.flush_cidx = -1;
qhp->wq.rq.msn = 1;
qhp->wq.rq.qid = resp.rqid;
qhp->wq.rq.size = resp.rq_size;
qhp->wq.rq.memsize = resp.rq_memsize;
if (ma_wr && resp.sq_memsize < (resp.sq_size + 1) *
sizeof *qhp->wq.sq.queue + 16*sizeof(__be64) ) {
ma_wr = 0;
fprintf(stderr, "libcxgb4 warning - downlevel iw_cxgb4 driver. "
"MA workaround disabled.\n");
}
ret = pthread_spin_init(&qhp->lock, PTHREAD_PROCESS_PRIVATE);
if (ret)
goto err3;
dbva = mmap(NULL, c4iw_page_size, PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.sq_db_gts_key);
if (dbva == MAP_FAILED)
goto err4;
qhp->wq.sq.udb = dbva;
if (!dev_is_t4(qhp->rhp)) {
unsigned long segment_offset = 128 * (qhp->wq.sq.qid &
qhp->wq.qid_mask);
if (segment_offset < c4iw_page_size) {
qhp->wq.sq.udb += segment_offset / 4;
qhp->wq.sq.wc_reg_available = 1;
} else
qhp->wq.sq.bar2_qid = qhp->wq.sq.qid & qhp->wq.qid_mask;
qhp->wq.sq.udb += 2;
}
qhp->wq.sq.queue = mmap(NULL, qhp->wq.sq.memsize,
PROT_READ|PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.sq_key);
if (qhp->wq.sq.queue == MAP_FAILED)
goto err5;
dbva = mmap(NULL, c4iw_page_size, PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.rq_db_gts_key);
if (dbva == MAP_FAILED)
goto err6;
qhp->wq.rq.udb = dbva;
if (!dev_is_t4(qhp->rhp)) {
unsigned long segment_offset = 128 * (qhp->wq.rq.qid &
qhp->wq.qid_mask);
if (segment_offset < c4iw_page_size) {
qhp->wq.rq.udb += segment_offset / 4;
qhp->wq.rq.wc_reg_available = 1;
} else
qhp->wq.rq.bar2_qid = qhp->wq.rq.qid & qhp->wq.qid_mask;
qhp->wq.rq.udb += 2;
}
qhp->wq.rq.queue = mmap(NULL, qhp->wq.rq.memsize,
PROT_READ|PROT_WRITE, MAP_SHARED,
pd->context->cmd_fd, resp.rq_key);
if (qhp->wq.rq.queue == MAP_FAILED)
goto err7;
qhp->wq.sq.sw_sq = calloc(qhp->wq.sq.size, sizeof (struct t4_swsqe));
if (!qhp->wq.sq.sw_sq)
goto err8;
qhp->wq.rq.sw_rq = calloc(qhp->wq.rq.size, sizeof (uint64_t));
if (!qhp->wq.rq.sw_rq)
goto err9;
if (t4_sq_onchip(&qhp->wq)) {
qhp->wq.sq.ma_sync = mmap(NULL, c4iw_page_size, PROT_WRITE,
MAP_SHARED, pd->context->cmd_fd,
resp.ma_sync_key);
if (qhp->wq.sq.ma_sync == MAP_FAILED)
goto err10;
qhp->wq.sq.ma_sync += (A_PCIE_MA_SYNC & (c4iw_page_size - 1));
}
if (ctx->status_page_size) {
qhp->wq.db_offp = &ctx->status_page->db_off;
} else {
qhp->wq.db_offp =
&qhp->wq.rq.queue[qhp->wq.rq.size].status.db_off;
}
PDBG("%s sq dbva %p sq qva %p sq depth %u sq memsize %lu "
" rq dbva %p rq qva %p rq depth %u rq memsize %lu\n",
__func__,
qhp->wq.sq.udb, qhp->wq.sq.queue,
qhp->wq.sq.size, qhp->wq.sq.memsize,
qhp->wq.rq.udb, qhp->wq.rq.queue,
qhp->wq.rq.size, qhp->wq.rq.memsize);
qhp->sq_sig_all = attr->sq_sig_all;
pthread_spin_lock(&dev->lock);
dev->qpid2ptr[qhp->wq.sq.qid] = qhp;
pthread_spin_unlock(&dev->lock);
INC_STAT(qp);
return &qhp->ibv_qp;
err10:
free(qhp->wq.rq.sw_rq);
err9:
free(qhp->wq.sq.sw_sq);
err8:
munmap((void *)qhp->wq.rq.queue, qhp->wq.rq.memsize);
err7:
munmap(MASKED(qhp->wq.rq.udb), c4iw_page_size);
err6:
munmap((void *)qhp->wq.sq.queue, qhp->wq.sq.memsize);
err5:
munmap(MASKED(qhp->wq.sq.udb), c4iw_page_size);
err4:
pthread_spin_destroy(&qhp->lock);
err3:
(void)ibv_cmd_destroy_qp(&qhp->ibv_qp);
err2:
free(qhp);
err1:
return NULL;
}
struct ibv_qp *c4iw_create_qp(struct ibv_pd *pd,
struct ibv_qp_init_attr *attr)
{
struct c4iw_dev *dev = to_c4iw_dev(pd->context->device);
if (dev->abi_version == 0)
return create_qp_v0(pd, attr);
return create_qp(pd, attr);
}
static void reset_qp(struct c4iw_qp *qhp)
{
PDBG("%s enter qp %p\n", __func__, qhp);
qhp->wq.sq.cidx = 0;
qhp->wq.sq.wq_pidx = qhp->wq.sq.pidx = qhp->wq.sq.in_use = 0;
qhp->wq.rq.cidx = qhp->wq.rq.pidx = qhp->wq.rq.in_use = 0;
qhp->wq.sq.oldest_read = NULL;
memset(qhp->wq.sq.queue, 0, qhp->wq.sq.memsize);
if (t4_sq_onchip(&qhp->wq))
mmio_flush_writes();
memset(qhp->wq.rq.queue, 0, qhp->wq.rq.memsize);
}
int c4iw_modify_qp(struct ibv_qp *ibqp, struct ibv_qp_attr *attr,
int attr_mask)
{
struct ibv_modify_qp cmd = {};
struct c4iw_qp *qhp = to_c4iw_qp(ibqp);
int ret;
PDBG("%s enter qp %p new state %d\n", __func__, ibqp, attr_mask & IBV_QP_STATE ? attr->qp_state : -1);
pthread_spin_lock(&qhp->lock);
if (t4_wq_in_error(&qhp->wq))
c4iw_flush_qp(qhp);
ret = ibv_cmd_modify_qp(ibqp, attr, attr_mask, &cmd, sizeof cmd);
if (!ret && (attr_mask & IBV_QP_STATE) && attr->qp_state == IBV_QPS_RESET)
reset_qp(qhp);
pthread_spin_unlock(&qhp->lock);
return ret;
}
int c4iw_destroy_qp(struct ibv_qp *ibqp)
{
int ret;
struct c4iw_qp *qhp = to_c4iw_qp(ibqp);
struct c4iw_dev *dev = to_c4iw_dev(ibqp->context->device);
PDBG("%s enter qp %p\n", __func__, ibqp);
pthread_spin_lock(&qhp->lock);
c4iw_flush_qp(qhp);
pthread_spin_unlock(&qhp->lock);
ret = ibv_cmd_destroy_qp(ibqp);
if (ret) {
return ret;
}
if (t4_sq_onchip(&qhp->wq)) {
qhp->wq.sq.ma_sync -= (A_PCIE_MA_SYNC & (c4iw_page_size - 1));
munmap((void *)qhp->wq.sq.ma_sync, c4iw_page_size);
}
munmap(MASKED(qhp->wq.sq.udb), c4iw_page_size);
munmap(MASKED(qhp->wq.rq.udb), c4iw_page_size);
munmap(qhp->wq.sq.queue, qhp->wq.sq.memsize);
munmap(qhp->wq.rq.queue, qhp->wq.rq.memsize);
pthread_spin_lock(&dev->lock);
dev->qpid2ptr[qhp->wq.sq.qid] = NULL;
pthread_spin_unlock(&dev->lock);
free(qhp->wq.rq.sw_rq);
free(qhp->wq.sq.sw_sq);
pthread_spin_destroy(&qhp->lock);
free(qhp);
return 0;
}
int c4iw_query_qp(struct ibv_qp *ibqp, struct ibv_qp_attr *attr,
int attr_mask, struct ibv_qp_init_attr *init_attr)
{
struct ibv_query_qp cmd;
struct c4iw_qp *qhp = to_c4iw_qp(ibqp);
int ret;
pthread_spin_lock(&qhp->lock);
if (t4_wq_in_error(&qhp->wq))
c4iw_flush_qp(qhp);
ret = ibv_cmd_query_qp(ibqp, attr, attr_mask, init_attr, &cmd, sizeof cmd);
pthread_spin_unlock(&qhp->lock);
return ret;
}
struct ibv_ah *c4iw_create_ah(struct ibv_pd *pd, struct ibv_ah_attr *attr)
{
return NULL;
}
int c4iw_destroy_ah(struct ibv_ah *ah)
{
return ENOSYS;
}
int c4iw_attach_mcast(struct ibv_qp *ibqp, const union ibv_gid *gid,
uint16_t lid)
{
struct c4iw_qp *qhp = to_c4iw_qp(ibqp);
int ret;
pthread_spin_lock(&qhp->lock);
if (t4_wq_in_error(&qhp->wq))
c4iw_flush_qp(qhp);
ret = ibv_cmd_attach_mcast(ibqp, gid, lid);
pthread_spin_unlock(&qhp->lock);
return ret;
}
int c4iw_detach_mcast(struct ibv_qp *ibqp, const union ibv_gid *gid,
uint16_t lid)
{
struct c4iw_qp *qhp = to_c4iw_qp(ibqp);
int ret;
pthread_spin_lock(&qhp->lock);
if (t4_wq_in_error(&qhp->wq))
c4iw_flush_qp(qhp);
ret = ibv_cmd_detach_mcast(ibqp, gid, lid);
pthread_spin_unlock(&qhp->lock);
return ret;
}
void c4iw_async_event(struct ibv_async_event *event)
{
PDBG("%s type %d obj %p\n", __func__, event->event_type,
event->element.cq);
switch (event->event_type) {
case IBV_EVENT_CQ_ERR:
break;
case IBV_EVENT_QP_FATAL:
case IBV_EVENT_QP_REQ_ERR:
case IBV_EVENT_QP_ACCESS_ERR:
case IBV_EVENT_PATH_MIG_ERR: {
struct c4iw_qp *qhp = to_c4iw_qp(event->element.qp);
pthread_spin_lock(&qhp->lock);
c4iw_flush_qp(qhp);
pthread_spin_unlock(&qhp->lock);
break;
}
case IBV_EVENT_SQ_DRAINED:
case IBV_EVENT_PATH_MIG:
case IBV_EVENT_COMM_EST:
case IBV_EVENT_QP_LAST_WQE_REACHED:
default:
break;
}
}