/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | qp.h | 13 int rvt_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, 15 int rvt_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 17 int rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata); 18 int rvt_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 20 int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 22 int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
|
H A D | mcast.h | 12 int rvt_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 13 int rvt_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
|
H A D | trace_qp.h | 21 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 26 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 27 __entry->qpn = qp->ibqp.qp_num; 51 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 58 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 59 __entry->qpn = qp->ibqp.qp_num;
|
H A D | trace_rc.h | 21 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 32 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 33 __entry->qpn = qp->ibqp.qp_num;
|
H A D | trace_tx.h | 52 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 72 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 75 __entry->qpn = qp->ibqp.qp_num; 76 __entry->qpt = qp->ibqp.qp_type; 119 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 131 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 134 __entry->qpn = qp->ibqp.qp_num; 135 __entry->qpt = qp->ibqp.qp_type;
|
H A D | qp.c | 426 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); 429 rvt_reset_qp(rdi, qp, qp->ibqp.qp_type); 582 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); 666 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); 692 if (qp->ibqp.qp_type == IB_QPT_SMI || 693 qp->ibqp.qp_type == IB_QPT_GSI) 714 ev.device = qp->ibqp.device; 715 ev.element.qp = &qp->ibqp; 717 qp->ibqp.event_handler(&ev, qp->ibqp 1029 rvt_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, struct ib_udata *udata) argument 1402 rvt_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 1677 rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata) argument 1718 rvt_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) argument 1781 rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 2143 rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 2509 struct ib_qp *ibqp = &qp->ibqp; local [all...] |
H A D | mcast.c | 232 * @ibqp: Infiniband qp 238 int rvt_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 240 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); 241 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); 247 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) 295 * @ibqp: Infiniband qp 301 int rvt_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 303 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); 304 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); 312 if (ibqp [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | wr.h | 104 int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, 106 int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 109 static inline int mlx5_ib_post_send_nodrain(struct ib_qp *ibqp, argument 113 return mlx5_ib_post_send(ibqp, wr, bad_wr, false); 116 static inline int mlx5_ib_post_send_drain(struct ib_qp *ibqp, argument 120 return mlx5_ib_post_send(ibqp, wr, bad_wr, true); 123 static inline int mlx5_ib_post_recv_nodrain(struct ib_qp *ibqp, argument 127 return mlx5_ib_post_recv(ibqp, wr, bad_wr, false); 130 static inline int mlx5_ib_post_recv_drain(struct ib_qp *ibqp, argument 134 return mlx5_ib_post_recv(ibqp, w [all...] |
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | qedr_roce_cm.h | 49 int qedr_gsi_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 51 int qedr_gsi_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
|
/linux-master/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.h | 49 struct ib_qp ibqp; member in struct:usnic_ib_qp_grp 105 struct usnic_ib_qp_grp *to_uqp_grp(struct ib_qp *ibqp) argument 107 return container_of(ibqp, struct usnic_ib_qp_grp, ibqp);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | qp.c | 163 struct ib_qp *ibqp = &qp->ibqp; local 164 struct hfi1_ibdev *dev = to_idev(ibqp->device); 169 sc = ah_to_sc(ibqp->device, &attr->ah_attr); 182 sc = ah_to_sc(ibqp->device, &attr->alt_ah_attr); 209 hfi1_update_ah_attr(qp->ibqp.device, &qp->remote_ah_attr); 217 ibp = to_iport(qp->ibqp.device, qp->port_num); 225 struct ib_qp *ibqp = &qp->ibqp; local 229 priv->s_sc = ah_to_sc(ibqp [all...] |
H A D | ud.c | 33 struct hfi1_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); 47 qp = rvt_lookup_qpn(ib_to_rvt(sqp->ibqp.device), &ibp->rvp, 55 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? 56 IB_QPT_UD : sqp->ibqp.qp_type; 57 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? 58 IB_QPT_UD : qp->ibqp.qp_type; 69 if (qp->ibqp.qp_num > 1) { 82 sqp->ibqp.qp_num, qp->ibqp.qp_num, 93 if (qp->ibqp [all...] |
H A D | trace_rc.h | 21 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 32 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 33 __entry->qpn = qp->ibqp.qp_num; 83 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 92 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 93 __entry->qpn = qp->ibqp.qp_num;
|
H A D | trace_misc.h | 58 TP_STRUCT__entry(DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 62 TP_fast_assign(DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 63 __entry->qpn = qp->ibqp.qp_num;
|
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 53 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); 73 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? 74 IB_QPT_UD : sqp->ibqp.qp_type; 75 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? 76 IB_QPT_UD : qp->ibqp.qp_type; 87 if (qp->ibqp.qp_num > 1) { 99 sqp->ibqp.qp_num, qp->ibqp.qp_num, 111 if (qp->ibqp.qp_num) { 149 if (qp->ibqp [all...] |
H A D | qib_ruc.c | 53 ev.device = qp->ibqp.device; 54 ev.element.qp = &qp->ibqp; 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); 114 0, qp->ibqp.qp_num, 153 0, qp->ibqp.qp_num, 210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); 266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); 271 if ((qp->ibqp.qp_type == IB_QPT_RC || 272 qp->ibqp [all...] |
/linux-master/drivers/infiniband/hw/mana/ |
H A D | qp.c | 91 static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd, argument 95 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); 262 static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd, argument 267 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); 401 int mana_ib_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *attr, argument 408 return mana_ib_create_qp_rss(ibqp, ibqp->pd, attr, 411 return mana_ib_create_qp_raw(ibqp, ibq 421 mana_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 476 mana_ib_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata) argument [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.h | 221 struct ib_qp ibqp; member in struct:erdma_qp 275 #define QP_ID(qp) ((qp)->ibqp.qp_num) 311 return container_of(qp, struct erdma_qp, ibqp); 339 int erdma_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *attr, 341 int erdma_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int mask, 343 int erdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int mask, 345 int erdma_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata); 354 void erdma_qp_get_ref(struct ib_qp *ibqp); 355 void erdma_qp_put_ref(struct ib_qp *ibqp); 357 int erdma_post_send(struct ib_qp *ibqp, cons [all...] |
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_qp.c | 61 *send_cq = to_vcq(qp->ibqp.send_cq); 62 *recv_cq = to_vcq(qp->ibqp.recv_cq); 185 * @ibqp: queue pair 191 int pvrdma_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, argument 194 struct pvrdma_qp *qp = to_vqp(ibqp); 195 struct pvrdma_dev *dev = to_vdev(ibqp->device); 234 init_attr->port_num > ibqp->device->phys_port_cnt) { 271 qp->rumem = ib_umem_get(ibqp->device, 284 qp->sumem = ib_umem_get(ibqp->device, ucmd.sbuf_addr, 302 ret = pvrdma_set_sq_size(to_vdev(ibqp 525 pvrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 671 pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 873 pvrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 966 pvrdma_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) argument [all...] |
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_loc.h | 38 int rxe_attach_mcast(struct ib_qp *ibqp, union ib_gid *mgid, u16 mlid); 39 int rxe_detach_mcast(struct ib_qp *ibqp, union ib_gid *mgid, u16 mlid); 118 return qp->ibqp.qp_num; 123 return qp->ibqp.qp_type; 133 if (qp->ibqp.qp_type == IB_QPT_RC || qp->ibqp.qp_type == IB_QPT_UC) 181 return rxe_wr_opcode_info[opcode].mask[qp->ibqp.qp_type];
|
H A D | rxe_comp.c | 221 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 404 wc->qp = &qp->ibqp; 408 uwc->qp_num = qp->ibqp.qp_num; 442 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 484 if (qp->ibqp.event_handler) { 487 ev.device = qp->ibqp.device; 488 ev.element.qp = &qp->ibqp; 490 qp->ibqp.event_handler(&ev, 491 qp->ibqp.qp_context); 552 ib_device_put(qp->ibqp [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_mcg.c | 120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 122 struct mthca_dev *dev = to_mdev(ibqp->device); 165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { 167 ibqp->qp_num); 171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); 214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 216 struct mthca_dev *dev = to_mdev(ibqp->device); 242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) 249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num);
|
H A D | mthca_provider.h | 254 struct ib_qp ibqp; member in struct:mthca_qp 311 static inline struct mthca_qp *to_mqp(struct ib_qp *ibqp) argument 313 return container_of(ibqp, struct mthca_qp, ibqp);
|
H A D | mthca_qp.c | 261 event.element.qp = &qp->ibqp; 262 if (qp->ibqp.event_handler) 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); 430 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, argument 433 struct mthca_dev *dev = to_mdev(ibqp->device); 434 struct mthca_qp *qp = to_mqp(ibqp); 556 static int __mthca_modify_qp(struct ib_qp *ibqp, argument 562 struct mthca_dev *dev = to_mdev(ibqp->device); 563 struct mthca_qp *qp = to_mqp(ibqp); 858 mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 1625 mthca_tavor_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 1823 mthca_tavor_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 1928 mthca_arbel_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 2162 mthca_arbel_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 109 event.element.qp = &qhp->ibqp; 110 if (qhp->ibqp.event_handler) 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); 153 c4iw_qp_add_ref(&qhp->ibqp); 217 c4iw_qp_rem_ref(&qhp->ibqp);
|