/linux-master/net/smc/ |
H A D | smc_core.h | 101 struct ib_qp_attr qp_attr; /* IB queue pair attributes */ member in struct:smc_link 499 lnk->qp_attr.cur_qp_state == IB_QPS_RTS;
|
H A D | smc_ib.c | 49 struct ib_qp_attr qp_attr; local 51 memset(&qp_attr, 0, sizeof(qp_attr)); 52 qp_attr.qp_state = IB_QPS_INIT; 53 qp_attr.pkey_index = 0; 54 qp_attr.port_num = lnk->ibport; 55 qp_attr.qp_access_flags = IB_ACCESS_LOCAL_WRITE 57 return ib_modify_qp(lnk->roce_qp, &qp_attr, 67 struct ib_qp_attr qp_attr; local 70 memset(&qp_attr, 97 struct ib_qp_attr qp_attr; local 116 struct ib_qp_attr qp_attr; local 666 struct ib_qp_init_attr qp_attr = { local [all...] |
H A D | smc_tx.c | 377 if (dst_len < link->qp_attr.cap.max_inline_data) {
|
/linux-master/drivers/nvme/target/ |
H A D | rdma.c | 1253 struct ib_qp_init_attr qp_attr = { }; local 1271 qp_attr.qp_context = queue; 1272 qp_attr.event_handler = nvmet_rdma_qp_event; 1273 qp_attr.send_cq = queue->cq; 1274 qp_attr.recv_cq = queue->cq; 1275 qp_attr.sq_sig_type = IB_SIGNAL_REQ_WR; 1276 qp_attr.qp_type = IB_QPT_RC; 1278 qp_attr.cap.max_send_wr = queue->send_queue_size + 1; 1281 qp_attr.cap.max_rdma_ctxs = queue->send_queue_size * factor; 1282 qp_attr [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | cm.c | 4099 struct ib_qp_attr *qp_attr, 4118 qp_attr->qp_access_flags = IB_ACCESS_REMOTE_WRITE; 4127 qp_attr->qp_access_flags |= IB_ACCESS_REMOTE_READ | 4131 qp_attr->pkey_index = cm_id_priv->av.pkey_index; 4133 qp_attr->port_num = cm_id_priv->av.port->port_num; 4146 struct ib_qp_attr *qp_attr, 4163 qp_attr->ah_attr = cm_id_priv->av.ah_attr; 4164 if ((qp_attr->ah_attr.type == RDMA_AH_ATTR_TYPE_IB) && 4167 qp_attr->ah_attr.ib.dlid = cm_id_priv->av.dlid_datapath; 4168 qp_attr 4098 cm_init_qp_init_attr(struct cm_id_private *cm_id_priv, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument 4145 cm_init_qp_rtr_attr(struct cm_id_private *cm_id_priv, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument 4198 cm_init_qp_rts_attr(struct cm_id_private *cm_id_priv, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument 4260 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
H A D | verbs.c | 1775 * If the user provided the qp_attr then we have to 2006 struct ib_qp_attr *qp_attr, 2009 return _ib_modify_qp(qp->real_qp, qp_attr, qp_attr_mask, NULL); 2014 struct ib_qp_attr *qp_attr, 2018 qp_attr->ah_attr.grh.sgid_attr = NULL; 2019 qp_attr->alt_ah_attr.grh.sgid_attr = NULL; 2022 qp->device->ops.query_qp(qp->real_qp, qp_attr, qp_attr_mask, 2005 ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask) argument 2013 ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument
|
H A D | ucma.c | 1218 struct ib_qp_attr qp_attr; local 1235 memset(&qp_attr, 0, sizeof qp_attr); 1236 qp_attr.qp_state = cmd.qp_state; 1238 ret = rdma_init_qp_attr(ctx->cm_id, &qp_attr, &resp.qp_attr_mask); 1243 ib_copy_qp_attr_to_user(ctx->cm_id->device, &resp, &qp_attr);
|
/linux-master/include/rdma/ |
H A D | ib_verbs.h | 2460 int (*modify_qp)(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 2462 int (*query_qp)(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 3754 * @qp_attr: On input, specifies the QP attributes to modify. On output, 3760 struct ib_qp_attr *qp_attr, 3767 * @qp_attr: The attributes of the specified QP. 3775 struct ib_qp_attr *qp_attr,
|
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 1159 struct ib_qp_attr qp_attr; local 1165 qp_attr.qp_state = IB_QPS_RTR; 1166 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask); 1170 qp_attr.max_dest_rd_atomic = 4; 1172 ret = ib_modify_qp(qp, &qp_attr, attr_mask); 1191 struct ib_qp_attr qp_attr; local 1195 qp_attr.qp_state = IB_QPS_RTS; 1196 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask); 1200 qp_attr.max_rd_atomic = 4; 1202 ret = ib_modify_qp(qp, &qp_attr, attr_mas 1214 struct ib_qp_attr qp_attr; local [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mlx5_ib.h | 1302 int mlx5_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 1490 int mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
|
H A D | qp.c | 4934 struct ib_qp_attr *qp_attr) 4954 qp_attr->sq_draining = 1; 4956 qp_attr->path_mtu = MLX5_GET(qpc, qpc, mtu); 4957 qp_attr->path_mig_state = to_ib_mig_state(MLX5_GET(qpc, qpc, pm_state)); 4958 qp_attr->qkey = MLX5_GET(qpc, qpc, q_key); 4959 qp_attr->rq_psn = MLX5_GET(qpc, qpc, next_rcv_psn); 4960 qp_attr->sq_psn = MLX5_GET(qpc, qpc, next_send_psn); 4961 qp_attr->dest_qp_num = MLX5_GET(qpc, qpc, remote_qpn); 4964 qp_attr->qp_access_flags |= IB_ACCESS_REMOTE_READ; 4966 qp_attr 4933 query_qp_attr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_attr *qp_attr) argument 4997 mlx5_ib_dct_query_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *mqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument 5062 mlx5_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 5356 static int hns_roce_v2_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, argument 5368 memset(qp_attr, 0, sizeof(*qp_attr)); 5374 qp_attr->qp_state = IB_QPS_RESET; 5394 qp_attr->qp_state = (enum ib_qp_state)hr_qp->state; 5395 qp_attr->path_mtu = (enum ib_mtu)hr_reg_read(&context, QPC_MTU); 5396 qp_attr->path_mig_state = IB_MIG_ARMED; 5397 qp_attr->ah_attr.type = RDMA_AH_ATTR_TYPE_ROCE; 5399 qp_attr->qkey = le32_to_cpu(context.qkey_xrcd); 5401 qp_attr [all...] |
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa.h | 150 int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, 184 int efa_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr,
|
H A D | efa_verbs.c | 302 int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, argument 323 memset(qp_attr, 0, sizeof(*qp_attr)); 331 qp_attr->qp_state = result.qp_state; 332 qp_attr->qkey = result.qkey; 333 qp_attr->sq_psn = result.sq_psn; 334 qp_attr->sq_draining = result.sq_draining; 335 qp_attr->port_num = 1; 336 qp_attr->rnr_retry = result.rnr_retry; 338 qp_attr 886 efa_modify_qp_validate(struct efa_dev *dev, struct efa_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) argument 930 efa_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_udata *udata) argument [all...] |
/linux-master/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 377 struct ib_qp_init_attr qp_attr; local 451 memset(&qp_attr, 0, sizeof qp_attr); 452 qp_attr.event_handler = qp_event_handler; 453 qp_attr.qp_context = &newxprt->sc_xprt; 454 qp_attr.port_num = newxprt->sc_port_num; 455 qp_attr.cap.max_rdma_ctxs = ctxts; 456 qp_attr.cap.max_send_wr = newxprt->sc_sq_depth - ctxts; 457 qp_attr.cap.max_recv_wr = rq_depth; 458 qp_attr [all...] |
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | verbs.c | 2742 struct ib_qp_attr *qp_attr, 2751 memset(qp_attr, 0, sizeof(*qp_attr)); 2758 qp_attr->qp_state = qedr_get_ibqp_state(params.state); 2760 qp_attr->qp_state = qedr_get_ibqp_state(QED_ROCE_QP_STATE_RTS); 2763 qp_attr->cur_qp_state = qedr_get_ibqp_state(params.state); 2764 qp_attr->path_mtu = ib_mtu_int_to_enum(params.mtu); 2765 qp_attr->path_mig_state = IB_MIG_MIGRATED; 2766 qp_attr->rq_psn = params.rq_psn; 2767 qp_attr 2741 qedr_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int attr_mask, struct ib_qp_init_attr *qp_init_attr) argument [all...] |
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | main.c | 1388 struct ib_qp_attr qp_attr; local 1391 qp_attr.qp_state = IB_QPS_ERR; 1402 bnxt_re_modify_qp(&qp->ib_qp, &qp_attr, mask,
|
H A D | ib_verbs.c | 1935 int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr, argument 1952 new_qp_state = qp_attr->qp_state; 1967 qp->qplib_qp.state = __from_ib_qp_state(qp_attr->qp_state); 1994 __from_ib_access_flags(qp_attr->qp_access_flags); 2003 qp->qplib_qp.pkey_index = qp_attr->pkey_index; 2007 qp->qplib_qp.qkey = qp_attr->qkey; 2011 rdma_ah_read_grh(&qp_attr->ah_attr); 2034 qp->qplib_qp.ah.sl = rdma_ah_get_sl(&qp_attr->ah_attr); 2036 qp_attr->ah_attr.roce.dmac); 2063 qp->qplib_qp.path_mtu = __from_ib_mtu(qp_attr 2180 bnxt_re_query_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument 2664 struct ib_qp_attr qp_attr; local [all...] |
H A D | ib_verbs.h | 214 int bnxt_re_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 216 int bnxt_re_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
|
/linux-master/fs/smb/server/ |
H A D | transport_rdma.c | 1873 struct ib_qp_init_attr qp_attr; local 1903 memset(&qp_attr, 0, sizeof(qp_attr)); 1904 qp_attr.event_handler = smb_direct_qpair_handler; 1905 qp_attr.qp_context = t; 1906 qp_attr.cap = *cap; 1907 qp_attr.sq_sig_type = IB_SIGNAL_REQ_WR; 1908 qp_attr.qp_type = IB_QPT_RC; 1909 qp_attr.send_cq = t->send_cq; 1910 qp_attr [all...] |
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_ib.c | 351 struct ib_qp_attr qp_attr; local 355 ret = ib_query_qp(priv->qp, &qp_attr, IB_QP_STATE, &query_init_attr); 362 __func__, priv->qp->qp_num, qp_attr.qp_state); 365 if (qp_attr.qp_state == IB_QPS_SQE) { 366 qp_attr.qp_state = IB_QPS_RTS; 368 ret = ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE); 378 priv->qp->qp_num, qp_attr.qp_state); 763 struct ib_qp_attr qp_attr; local 767 ret = ib_query_qp(qp, &qp_attr, IB_QP_STATE, &query_init_attr); 773 if (new_state == IB_QPS_ERR && qp_attr 799 struct ib_qp_attr qp_attr; local [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.c | 504 int siw_query_qp(struct ib_qp *base_qp, struct ib_qp_attr *qp_attr, argument 510 if (base_qp && qp_attr && qp_init_attr) { 516 qp_attr->qp_state = siw_qp_state_to_ib_qp_state[qp->attrs.state]; 517 qp_attr->cap.max_inline_data = SIW_MAX_INLINE; 518 qp_attr->cap.max_send_wr = qp->attrs.sq_size; 519 qp_attr->cap.max_send_sge = qp->attrs.sq_max_sges; 520 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; 521 qp_attr->cap.max_recv_sge = qp->attrs.rq_max_sges; 522 qp_attr->path_mtu = ib_mtu_int_to_enum(sdev->netdev->mtu); 523 qp_attr [all...] |
H A D | siw_verbs.h | 55 int siw_query_qp(struct ib_qp *base_qp, struct ib_qp_attr *qp_attr,
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 1543 int erdma_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, argument 1549 if (ibqp && qp_attr && qp_init_attr) { 1556 qp_attr->cap.max_inline_data = ERDMA_MAX_INLINE; 1559 qp_attr->cap.max_send_wr = qp->attrs.sq_size; 1560 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; 1561 qp_attr->cap.max_send_sge = qp->attrs.max_send_sge; 1562 qp_attr->cap.max_recv_sge = qp->attrs.max_recv_sge; 1564 qp_attr->path_mtu = ib_mtu_int_to_enum(dev->netdev->mtu); 1565 qp_attr->max_rd_atomic = qp->attrs.irq_size; 1566 qp_attr [all...] |
/linux-master/fs/smb/client/ |
H A D | smbdirect.c | 1494 struct ib_qp_init_attr qp_attr; local 1564 memset(&qp_attr, 0, sizeof(qp_attr)); 1565 qp_attr.event_handler = smbd_qp_async_error_upcall; 1566 qp_attr.qp_context = info; 1567 qp_attr.cap.max_send_wr = info->send_credit_target; 1568 qp_attr.cap.max_recv_wr = info->receive_credit_max; 1569 qp_attr.cap.max_send_sge = SMBDIRECT_MAX_SEND_SGE; 1570 qp_attr.cap.max_recv_sge = SMBDIRECT_MAX_RECV_SGE; 1571 qp_attr [all...] |