/linux-master/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_verbs.h | 46 int usnic_ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 53 int usnic_ib_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *init_attr, 55 int usnic_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
|
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_driver.c | 301 struct rvt_qp *qp = NULL; local 346 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 347 if (!qp) { 356 spin_lock(&qp->r_lock); 359 if (!(ib_rvt_state_ops[qp->state] & 365 switch (qp->ibqp.qp_type) { 371 qp, 379 diff = qib_cmp24(psn, qp->r_psn); 380 if (!qp->r_nak_state && diff >= 0) { 382 qp 448 struct rvt_qp *qp, *nqp; local [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | verbs.c | 21 #include "qp.h" 101 static int pio_wait(struct rvt_qp *qp, 276 if (!(ib_rvt_state_ops[packet->qp->state] & RVT_PROCESS_RECV_OK)) 279 packet->qp->allowed_ops) || 286 static u64 hfi1_fault_tx(struct rvt_qp *qp, u8 opcode, u64 pbc) argument 319 if (packet->qp->ibqp.qp_type != IB_QPT_RC || 320 !(ib_rvt_state_ops[packet->qp->state] & RVT_PROCESS_RECV_OK)) 360 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); 361 if (!packet->qp) 363 spin_lock_irqsave(&packet->qp 559 struct rvt_qp *qp = NULL; local 591 struct rvt_qp *qp = tx->qp; local 609 hfi1_wait_kmem(struct rvt_qp *qp) argument 627 wait_kmem(struct hfi1_ibdev *dev, struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument 697 update_tx_opstats(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u32 plen) argument 797 hfi1_verbs_send_dma(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc) argument 881 pio_wait(struct rvt_qp *qp, struct send_context *sc, struct hfi1_pkt_state *ps, u32 flag) argument 930 struct rvt_qp *qp = (struct rvt_qp *)arg; local 937 hfi1_verbs_send_pio(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc) argument 1173 get_send_routine(struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument 1213 hfi1_verbs_send(struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument 1917 struct rvt_qp *qp = packet->qp; local [all...] |
H A D | trace_misc.h | 56 TP_PROTO(struct rvt_qp *qp, u8 opcode), 57 TP_ARGS(qp, opcode), 58 TP_STRUCT__entry(DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 62 TP_fast_assign(DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 63 __entry->qpn = qp->ibqp.qp_num;
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | qpc.c | 7 #include <linux/mlx5/qp.h> 10 #include "qp.h" 114 struct mlx5_core_qp *qp; local 146 qp = (struct mlx5_core_qp *)common; 147 qp->event(qp, event_type); 160 struct mlx5_core_qp *qp, int rsc_type) 165 qp->common.res = rsc_type; 168 qp->qpn | (rsc_type << MLX5_USER_INDEX_LEN), 169 qp); 159 create_resource_common(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, int rsc_type) argument 181 destroy_resource_common(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp) argument 199 struct mlx5_core_qp *qp = &dct->mqp; local 210 struct mlx5_core_qp *qp = &dct->mqp; local 232 mlx5_qpc_create_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, u32 *in, int inlen, u32 *out) argument 268 struct mlx5_core_qp *qp = &dct->mqp; local 306 mlx5_core_destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp) argument 479 mlx5_core_qp_modify(struct mlx5_ib_dev *dev, u16 opcode, u32 opt_param_mask, void *qpc, struct mlx5_core_qp *qp, u32 *ece) argument 523 mlx5_core_qp_query(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, u32 *out, int outlen, bool qpc_ext) argument 539 struct mlx5_core_qp *qp = &dct->mqp; local [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_cm.c | 270 void erdma_qp_cm_drop(struct erdma_qp *qp) argument 272 struct erdma_cep *cep = qp->cep; 274 if (!qp->cep) 313 if (cep->qp) { 314 cep->qp = NULL; 315 erdma_qp_put(qp); 571 struct erdma_qp *qp = cep->qp; local 605 down_write(&qp->state_lock); 606 if (qp 858 struct erdma_qp *qp = cep->qp; local 1003 struct erdma_qp *qp; local 1131 struct erdma_qp *qp; local [all...] |
H A D | erdma_cq.c | 116 struct erdma_qp *qp; local 135 qp = find_qp_by_qpn(dev, qpn); 136 if (!qp) 139 kern_qp = &qp->kern_qp; 147 depth = qp->attrs.sq_size; 148 wqe_hdr = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, 149 qp->attrs.sq_size, SQEBB_SHIFT); 155 depth = qp->attrs.rq_size; 176 wc->qp = &qp [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | user.h | 258 int irdma_uk_inline_rdma_write(struct irdma_qp_uk *qp, 260 int irdma_uk_inline_send(struct irdma_qp_uk *qp, 262 int irdma_uk_post_nop(struct irdma_qp_uk *qp, u64 wr_id, bool signaled, 264 int irdma_uk_post_receive(struct irdma_qp_uk *qp, 266 void irdma_uk_qp_post_wr(struct irdma_qp_uk *qp); 267 int irdma_uk_rdma_read(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info, 269 int irdma_uk_rdma_write(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info, 271 int irdma_uk_send(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info, 273 int irdma_uk_stag_local_invalidate(struct irdma_qp_uk *qp, 292 void irdma_uk_cq_set_resized_cnt(struct irdma_cq_uk *qp, u1 [all...] |
H A D | ctrl.c | 13 * irdma_get_qp_from_list - get next qp from a list 14 * @head: Listhead of qp's 15 * @qp: current qp 18 struct irdma_sc_qp *qp) 26 if (!qp) { 29 lastentry = &qp->list; 39 * irdma_sc_suspend_resume_qps - suspend/resume all qp's on VSI 45 struct irdma_sc_qp *qp = NULL; local 50 qp 17 irdma_get_qp_from_list(struct list_head *head, struct irdma_sc_qp *qp) argument 123 irdma_qp_rem_qos(struct irdma_sc_qp *qp) argument 143 irdma_qp_add_qos(struct irdma_sc_qp *qp) argument 375 irdma_sc_qp_init(struct irdma_sc_qp *qp, struct irdma_qp_init_info *info) argument 448 irdma_sc_qp_create(struct irdma_sc_qp *qp, struct irdma_create_qp_info *info, u64 scratch, bool post_sq) argument 499 irdma_sc_qp_modify(struct irdma_sc_qp *qp, struct irdma_modify_qp_info *info, u64 scratch, bool post_sq) argument 569 irdma_sc_qp_destroy(struct irdma_sc_qp *qp, u64 scratch, bool remove_hash_idx, bool ignore_mw_bnd, bool post_sq) argument 637 irdma_sc_qp_setctx_roce(struct irdma_sc_qp *qp, __le64 *qp_ctx, struct irdma_qp_host_ctx_info *info) argument 869 irdma_sc_qp_setctx(struct irdma_sc_qp *qp, __le64 *qp_ctx, struct irdma_qp_host_ctx_info *info) argument 1291 irdma_sc_mr_fast_register(struct irdma_sc_qp *qp, struct irdma_fast_reg_stag_info *info, bool post_sq) argument 1364 irdma_sc_gen_rts_ae(struct irdma_sc_qp *qp) argument 1400 irdma_sc_send_lsmm(struct irdma_sc_qp *qp, void *lsmm_buf, u32 size, irdma_stag stag) argument 1443 irdma_sc_send_rtt(struct irdma_sc_qp *qp, bool read) argument 1530 irdma_bld_termhdr_ctrl(struct irdma_sc_qp *qp, struct irdma_terminate_hdr *hdr, enum irdma_flush_opcode opcode, u8 layer_etype, u8 err) argument 1581 irdma_bld_terminate_hdr(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument 1759 irdma_terminate_send_fin(struct irdma_sc_qp *qp) argument 1770 irdma_terminate_connection(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument 1790 irdma_terminate_received(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument 2215 irdma_sc_qp_flush_wqes(struct irdma_sc_qp *qp, struct irdma_qp_flush_info *info, u64 scratch, bool post_sq) argument 2289 irdma_sc_gen_ae(struct irdma_sc_qp *qp, struct irdma_gen_ae_info *info, u64 scratch, bool post_sq) argument 2408 irdma_sc_suspend_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) argument 2438 irdma_sc_resume_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) argument [all...] |
H A D | puda.h | 30 struct irdma_qp_uk *qp; member in struct:irdma_puda_cmpl_info 101 struct irdma_sc_qp qp; member in struct:irdma_puda_rsrc 154 int irdma_puda_send(struct irdma_sc_qp *qp, struct irdma_puda_send_info *info); 168 void irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp); 171 int irdma_cqp_qp_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp); 173 int irdma_cqp_qp_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp); 175 void irdma_puda_ieq_get_ah_info(struct irdma_sc_qp *qp, 182 void irdma_ieq_process_fpdus(struct irdma_sc_qp *qp, 184 void irdma_ieq_cleanup_qp(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp);
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 245 static int dump_qp(unsigned long id, struct c4iw_qp *qp, argument 250 if (id != qp->wq.sq.qid) 257 if (qp->ep) { 258 struct c4iw_ep *ep = qp->ep; 268 "rc qp sq id %u %s id %u state %u " 271 qp->wq.sq.qid, qp->srq ? "srq" : "rq", 272 qp->srq ? qp->srq->idx : qp 328 struct c4iw_qp *qp; local 1270 struct c4iw_qp *qp; local 1285 resume_rc_qp(struct c4iw_qp *qp) argument 1298 struct c4iw_qp *qp; local 1318 struct c4iw_qp *qp; local 1374 struct c4iw_qp *qp = qp_list->qps[idx]; local 1417 struct c4iw_qp *qp; local [all...] |
H A D | Makefile | 7 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o \
|
/linux-master/drivers/infiniband/hw/mana/ |
H A D | qp.c | 95 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local 120 "Failed copy from udata for create rss-qp, err %d\n", 158 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n", 174 qp->port = port; 237 "Failed to copy to udata create rss-qp, %d\n", 267 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local 292 "Failed to copy from udata create qp-raw, %d\n", err); 313 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n", 324 qp->port = port; 329 err = mana_ib_create_queue(mdev, ucmd.sq_buf_addr, ucmd.sq_buf_size, &qp 428 mana_ib_destroy_qp_rss(struct mana_ib_qp *qp, struct ib_rwq_ind_table *ind_tbl, struct ib_udata *udata) argument 454 mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata) argument 478 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local [all...] |
/linux-master/lib/crypto/mpi/ |
H A D | mpi-div.c | 86 mpi_ptr_t qp, rp; local 133 qp = quot->d; 134 rlimb = mpihelp_divmod_1(qp, np, nsize, dp[0]); 135 qsize -= qp[qsize - 1] == 0; 149 qp = quot->d; 153 if (qp == np) { /* Copy NP object to temporary space. */ 155 MPN_COPY(np, qp, nsize); 158 qp = rp + dsize; 192 if (dp == rp || (quot && (dp == qp))) { 207 q_limb = mpihelp_divrem(qp, [all...] |
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 357 static int ehea_qp_alloc_register(struct ehea_qp *qp, struct hw_queue *hw_queue, argument 378 qp->fw_handle, rpage, 1); 402 struct ehea_qp *qp; local 407 qp = kzalloc(sizeof(*qp), GFP_KERNEL); 408 if (!qp) 411 qp->adapter = adapter; 414 &qp->fw_handle, &qp->epas); 425 ret = ehea_qp_alloc_register(qp, 490 ehea_destroy_qp_res(struct ehea_qp *qp, u64 force) argument 513 ehea_destroy_qp(struct ehea_qp *qp) argument [all...] |
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | qedr_iw_cm.c | 84 struct qedr_qp *qp = container_of(ref, struct qedr_qp, refcnt); local 86 complete(&qp->qp_rel_comp); 94 if (ep->qp) 95 kref_put(&ep->qp->refcnt, qedr_iw_free_qp); 188 struct ib_qp *ibqp = &ep->qp->ibqp; 196 event.element.qp = ibqp; 215 struct qedr_qp *qp = ep->qp; local 218 /* The qp won't be released until we release the ep. 220 * function, therefore it is safe to access qp 521 struct qedr_qp *qp; local 543 struct qedr_qp *qp; local 741 struct qedr_qp *qp; local 802 struct qedr_qp *qp = get_qedr_qp(ibqp); local 809 struct qedr_qp *qp = get_qedr_qp(ibqp); local [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_rdma.c | 157 p_rdma_info->num_qps = num_cons / 2; /* 2 cids per qp */ 1191 struct qed_rdma_qp *qp, 1197 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "icid = %08x\n", qp->icid); 1199 /* The following fields are filled in from qp and not FW as they can't 1202 out_params->mtu = qp->mtu; 1203 out_params->dest_qp = qp->dest_qp; 1204 out_params->incoming_atomic_en = qp->incoming_atomic_en; 1205 out_params->e2e_flow_control_en = qp->e2e_flow_control_en; 1206 out_params->incoming_rdma_read_en = qp->incoming_rdma_read_en; 1207 out_params->incoming_rdma_write_en = qp 1190 qed_rdma_query_qp(void *rdma_cxt, struct qed_rdma_qp *qp, struct qed_rdma_query_qp_out_params *out_params) argument 1230 qed_rdma_destroy_qp(void *rdma_cxt, struct qed_rdma_qp *qp) argument 1255 struct qed_rdma_qp *qp; local 1345 qed_rdma_modify_qp(void *rdma_cxt, struct qed_rdma_qp *qp, struct qed_rdma_modify_qp_in_params *params) argument [all...] |
H A D | qed_iwarp.h | 114 struct qed_rdma_qp *qp; member in struct:qed_iwarp_ep 176 struct qed_rdma_qp *qp, 179 int qed_iwarp_modify_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp, 182 int qed_iwarp_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp); 184 int qed_iwarp_fw_destroy(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp); 186 void qed_iwarp_query_qp(struct qed_rdma_qp *qp,
|
/linux-master/drivers/crypto/hisilicon/ |
H A D | qm.c | 845 static void qm_cq_head_update(struct hisi_qp *qp) argument 847 if (qp->qp_status.cq_head == qp->cq_depth - 1) { 848 qp->qp_status.cqc_phase = !qp->qp_status.cqc_phase; 849 qp->qp_status.cq_head = 0; 851 qp->qp_status.cq_head++; 855 static void qm_poll_req_cb(struct hisi_qp *qp) argument 857 struct qm_cqe *cqe = qp->cqe + qp 883 struct hisi_qp *qp; local 972 qm_set_qp_disable(struct hisi_qp *qp, int offset) argument 988 struct hisi_qp *qp = &qm->qp_array[qp_id]; local 1071 qm_init_qp_status(struct hisi_qp *qp) argument 1692 qm_stop_qp(struct hisi_qp *qp) argument 1808 qm_get_avail_sqe(struct hisi_qp *qp) argument 1819 hisi_qm_unset_hw_reset(struct hisi_qp *qp) argument 1831 struct hisi_qp *qp; local 1877 struct hisi_qp *qp; local 1900 hisi_qm_release_qp(struct hisi_qp *qp) argument 1914 qm_sq_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument 1940 qm_cq_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument 1969 qm_qp_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument 1982 qm_start_qp_nolock(struct hisi_qp *qp, unsigned long arg) argument 2013 hisi_qm_start_qp(struct hisi_qp *qp, unsigned long arg) argument 2032 qp_stop_fail_cb(struct hisi_qp *qp) argument 2095 qm_drain_qp(struct hisi_qp *qp) argument 2130 qm_stop_qp_nolock(struct hisi_qp *qp) argument 2169 hisi_qm_stop_qp(struct hisi_qp *qp) argument 2192 hisi_qp_send(struct hisi_qp *qp, const void *msg) argument 2233 qm_qp_event_notifier(struct hisi_qp *qp) argument 2264 struct hisi_qp *qp; local 2283 struct hisi_qp *qp = q->priv; local 2293 struct hisi_qp *qp = q->priv; local 2344 struct hisi_qp *qp = q->priv; local 2351 struct hisi_qp *qp = q->priv; local 2380 struct hisi_qp *qp = q->priv; local 2398 struct hisi_qp *qp = q->priv; local 2408 struct hisi_qp *qp = q->priv; local 2746 struct hisi_qp *qp; local 3089 struct hisi_qp *qp; local 3119 struct hisi_qp *qp; local 3140 struct hisi_qp *qp; local [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | restrack.c | 195 struct ib_qp *qp = container_of(res, struct ib_qp, res); local 197 WARN_ONCE(qp->qp_num >> 24 || qp->port >> 8, 198 "QP number 0x%0X and port 0x%0X", qp->qp_num, 199 qp->port); 200 res->id = qp->qp_num; 201 if (qp->qp_type == IB_QPT_SMI || qp->qp_type == IB_QPT_GSI) 202 res->id |= qp->port << 24; 207 if (qp [all...] |
H A D | uverbs_std_types_qp.c | 15 struct ib_qp *qp = uobject->object; local 30 } else if (qp == qp->real_qp) { 31 ib_uverbs_detach_umcast(qp, uqp); 34 ret = ib_destroy_qp_user(qp, &attrs->driver_udata); 91 struct ib_qp *qp; local 251 qp = ib_create_qp_user(device, pd, &attr, &attrs->driver_udata, obj, 253 if (IS_ERR(qp)) { 254 ret = PTR_ERR(qp); 257 ib_qp_usecnt_inc(qp); [all...] |
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 665 struct ocrdma_qp *qp) 670 if (qp == NULL) 672 ocrdma_qp_state_change(qp, new_ib_qps, &old_ib_qps); 678 struct ocrdma_qp *qp = NULL; local 691 * Some FW version returns wrong qp or cq ids in CQEs. 697 qp = dev->qp_tbl[qpid]; 698 if (qp == NULL) { 733 ib_evt.element.qp = &qp->ibqp; 735 ocrdma_process_qpcat_error(dev, qp); 664 ocrdma_process_qpcat_error(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument 905 struct ocrdma_qp *qp; local 2078 ocrdma_is_qp_in_sq_flushlist(struct ocrdma_cq *cq, struct ocrdma_qp *qp) argument 2091 ocrdma_is_qp_in_rq_flushlist(struct ocrdma_cq *cq, struct ocrdma_qp *qp) argument 2104 ocrdma_flush_qp(struct ocrdma_qp *qp) argument 2122 ocrdma_init_hwq_ptr(struct ocrdma_qp *qp) argument 2130 ocrdma_qp_state_change(struct ocrdma_qp *qp, enum ib_qp_state new_ib_state, enum ib_qp_state *old_ib_state) argument 2161 ocrdma_set_create_qp_mbx_access_flags(struct ocrdma_qp *qp) argument 2177 ocrdma_set_create_qp_sq_cmd(struct ocrdma_create_qp_req *cmd, struct ib_qp_init_attr *attrs, struct ocrdma_qp *qp) argument 2232 ocrdma_set_create_qp_rq_cmd(struct ocrdma_create_qp_req *cmd, struct ib_qp_init_attr *attrs, struct ocrdma_qp *qp) argument 2279 ocrdma_set_create_qp_dpp_cmd(struct ocrdma_create_qp_req *cmd, struct ocrdma_pd *pd, struct ocrdma_qp *qp, u8 enable_dpp_cq, u16 dpp_cq_id) argument 2295 ocrdma_set_create_qp_ird_cmd(struct ocrdma_create_qp_req *cmd, struct ocrdma_qp *qp) argument 2328 ocrdma_get_create_qp_rsp(struct ocrdma_create_qp_rsp *rsp, struct ocrdma_qp *qp, struct ib_qp_init_attr *attrs, u16 *dpp_offset, u16 *dpp_credit_lmt) argument 2363 ocrdma_mbx_create_qp(struct ocrdma_qp *qp, struct ib_qp_init_attr *attrs, u8 enable_dpp_cq, u16 dpp_cq_id, u16 *dpp_offset, u16 *dpp_credit_lmt) argument 2462 ocrdma_mbx_query_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ocrdma_qp_params *param) argument 2483 ocrdma_set_av_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) argument 2569 ocrdma_set_qp_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) argument 2678 ocrdma_mbx_modify_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ib_qp_attr *attrs, int attr_mask) argument 2714 ocrdma_mbx_destroy_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument [all...] |
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_verbs.c | 61 ret = ib_modify_qp(priv->qp, qp_attr, IB_QP_QKEY); 69 ret = ib_attach_mcast(priv->qp, mgid, mlid); 84 ret = ib_detach_mcast(priv->qp, mgid, mlid); 108 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); 117 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); 127 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); 137 if (ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE)) 212 priv->qp = ib_create_qp(priv->pd, &init_attr); 213 if (IS_ERR(priv->qp)) { 259 if (priv->qp) { [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_resources.c | 36 #include <linux/mlx4/qp.h> 78 /* disable multicast loopback to qp with same counter */ 89 en_dbg(HW, priv, "Setting RX qp %x tunnel mode to RX tunneled & non-tunneled\n", qpn); 94 int mlx4_en_change_mcast_lb(struct mlx4_en_priv *priv, struct mlx4_qp *qp, argument 104 ret = mlx4_update_qp(priv->mdev->dev, qp->qpn, 111 void mlx4_en_sqp_event(struct mlx4_qp *qp, enum mlx4_event event) argument
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe.h | 52 #define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device, \ 53 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) 73 #define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \ 74 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) 94 #define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp) [all...] |