/linux-master/drivers/interconnect/qcom/ |
H A D | osm-l3.c | 128 struct qcom_osm_l3_icc_provider *qp; local 136 qp = to_osm_l3_provider(provider); 141 for (index = 0; index < qp->max_state - 1; index++) { 142 if (qp->lut_tables[index] >= rate) 146 writel_relaxed(index, qp->base + qp->reg_perf_state); 153 struct qcom_osm_l3_icc_provider *qp = platform_get_drvdata(pdev); local 155 icc_provider_deregister(&qp->provider); 156 icc_nodes_remove(&qp->provider); 163 struct qcom_osm_l3_icc_provider *qp; local [all...] |
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_net.c | 23 static struct dst_entry *rxe_find_route4(struct rxe_qp *qp, argument 39 rxe_dbg_qp(qp, "no route to %pI4\n", &daddr->s_addr); 47 static struct dst_entry *rxe_find_route6(struct rxe_qp *qp, argument 65 rxe_dbg_qp(qp, "no route to %pI6\n", daddr); 70 rxe_dbg_qp(qp, "no route to %pI6\n", daddr); 82 static struct dst_entry *rxe_find_route6(struct rxe_qp *qp, argument 93 struct rxe_qp *qp, 98 if (qp_type(qp) == IB_QPT_RC) 99 dst = sk_dst_get(qp->sk->sk); 101 if (!dst || !dst_check(dst, qp 92 rxe_find_route(struct net_device *ndev, struct rxe_qp *qp, struct rxe_av *av) argument 282 struct rxe_qp *qp = pkt->qp; local 308 struct rxe_qp *qp = pkt->qp; local 351 struct rxe_qp *qp; local 428 rxe_xmit_packet(struct rxe_qp *qp, struct rxe_pkt_info *pkt, struct sk_buff *skb) argument [all...] |
H A D | rxe_verbs.c | 526 /* qp */ 532 struct rxe_qp *qp = to_rqp(ibqp); local 549 qp->is_user = true; 552 qp->is_user = false; 567 err = rxe_add_to_pool(&rxe->qp_pool, qp); 569 rxe_dbg_dev(rxe, "unable to create qp, err = %d\n", err); 573 err = rxe_qp_from_init(rxe, qp, pd, init, uresp, ibqp->pd, udata); 575 rxe_dbg_qp(qp, "create qp failed, err = %d\n", err); 579 rxe_finalize(qp); 595 struct rxe_qp *qp = to_rqp(ibqp); local 632 struct rxe_qp *qp = to_rqp(ibqp); local 642 struct rxe_qp *qp = to_rqp(ibqp); local 665 validate_send_wr(struct rxe_qp *qp, const struct ib_send_wr *ibwr, unsigned int *maskp, unsigned int *lengthp) argument 726 init_send_wr(struct rxe_qp *qp, struct rxe_send_wr *wr, const struct ib_send_wr *ibwr) argument 820 init_send_wqe(struct rxe_qp *qp, const struct ib_send_wr *ibwr, unsigned int mask, unsigned int length, struct rxe_send_wqe *wqe) argument 858 post_one_send(struct rxe_qp *qp, const struct ib_send_wr *ibwr) argument 885 rxe_post_send_kernel(struct rxe_qp *qp, const struct ib_send_wr *ibwr, const struct ib_send_wr **bad_wr) argument 916 struct rxe_qp *qp = to_rqp(ibqp); local 1006 struct rxe_qp *qp = to_rqp(ibqp); local [all...] |
H A D | rxe_mcast.c | 11 * struct rxe_mca ('mca'). An mcg is allocated each time a qp is 15 * qp is attached to the same mgid. It is cleaned up when the last qp 16 * is detached from the mcg. Each time a qp is attached to an mcg an 17 * mca is created. It holds a pointer to the qp and is added to a list 18 * of qp's that are attached to the mcg. The qp_list is used to replicate 254 * no qp's are attached to mcg 271 * Context: no qp's are attached to mcg 285 * @qp: qp objec 294 __rxe_init_mca(struct rxe_qp *qp, struct rxe_mcg *mcg, struct rxe_mca *mca) argument 331 rxe_attach_mcg(struct rxe_mcg *mcg, struct rxe_qp *qp) argument 396 rxe_detach_mcg(struct rxe_mcg *mcg, struct rxe_qp *qp) argument 437 struct rxe_qp *qp = to_rqp(ibqp); local 467 struct rxe_qp *qp = to_rqp(ibqp); local [all...] |
H A D | rxe_task.c | 28 * a qp reference to cover the gap from now until the task finishes. 38 WARN_ON(rxe_read(task->qp) <= 0); 41 rxe_get(task->qp); 111 WARN_ON(rxe_read(task->qp) <= 0); 115 rxe_put(task->qp); 127 ret = task->func(task->qp); 159 rxe_dbg_qp(task->qp, "unexpected task state = %d\n", 169 task->qp, 181 rxe_put(task->qp); 190 int rxe_init_task(struct rxe_task *task, struct rxe_qp *qp, argument [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | trace_tx.h | 49 TP_PROTO(struct rvt_qp *qp, struct rvt_swqe *wqe, int wr_num_sge), 50 TP_ARGS(qp, wqe, wr_num_sge), 52 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 72 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 75 __entry->qpn = qp->ibqp.qp_num; 76 __entry->qpt = qp->ibqp.qp_type; 81 __entry->size = qp->s_size; 82 __entry->avail = qp->s_avail; 83 __entry->head = qp->s_head; 84 __entry->last = qp [all...] |
H A D | mcast.c | 31 * @qp: the QP to link 33 static struct rvt_mcast_qp *rvt_mcast_qp_alloc(struct rvt_qp *qp) argument 41 mqp->qp = qp; 42 rvt_get_qp(qp); 50 struct rvt_qp *qp = mqp->qp; local 53 rvt_put_qp(qp); 186 if (p->qp == mqp->qp) { 240 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); local 303 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); local [all...] |
H A D | Makefile | 11 rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
|
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.c | 429 struct bnxt_re_qp *qp = container_of(qplib_qp, struct bnxt_re_qp, local 431 struct ib_pd *ib_pd = qp->ib_qp.pd; 442 ibdev_dbg(&qp->rdev->ibdev, 444 wqe.bind.r_key, qp->qplib_qp.id, pd); 445 rc = bnxt_qplib_post_send(&qp->qplib_qp, &wqe); 447 ibdev_err(&qp->rdev->ibdev, "Failed to bind fence-WQE\n"); 450 bnxt_qplib_post_send_db(&qp->qplib_qp); 819 unsigned long bnxt_re_lock_cqs(struct bnxt_re_qp *qp) 820 __acquires(&qp->scq->cq_lock) __acquires(&qp 844 bnxt_re_destroy_gsi_sqp(struct bnxt_re_qp *qp) argument 891 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 980 bnxt_re_setup_swqe_size(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr) argument 1019 bnxt_re_init_user_qp(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd, struct bnxt_re_qp *qp, struct ib_udata *udata) argument 1137 struct bnxt_re_qp *qp; local 1198 bnxt_re_init_rq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument 1239 bnxt_re_adjust_gsi_rq_attr(struct bnxt_re_qp *qp) argument 1257 bnxt_re_init_sq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument 1303 bnxt_re_adjust_gsi_sq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument 1350 bnxt_re_init_qp_attr(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd, struct ib_qp_init_attr *init_attr, struct ib_udata *udata) argument 1426 bnxt_re_create_shadow_gsi(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd) argument 1471 bnxt_re_create_gsi_qp(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd, struct ib_qp_init_attr *init_attr) argument 1526 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 1907 struct bnxt_re_qp *qp = rdev->gsi_ctx.gsi_sqp; local 1938 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 2183 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 2241 bnxt_re_build_qp1_send_v2(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument 2415 bnxt_re_build_qp1_shadow_qp_recv(struct bnxt_re_qp *qp, const struct ib_recv_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument 2458 is_ud_qp(struct bnxt_re_qp *qp) argument 2464 bnxt_re_build_send_wqe(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2657 bnxt_ud_qp_hw_stall_workaround(struct bnxt_re_qp *qp) argument 2673 bnxt_re_post_send_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_send_wr *wr) argument 2735 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 2833 bnxt_re_post_recv_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_recv_wr *wr) argument 2870 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local 3609 bnxt_re_process_res_ud_wc(struct bnxt_re_qp *qp, struct ib_wc *wc, struct bnxt_qplib_cqe *cqe) argument 3644 send_phantom_wqe(struct bnxt_re_qp *qp) argument 3669 struct bnxt_re_qp *qp, *sh_qp; local [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | puda.c | 14 static void irdma_ilq_putback_rcvbuf(struct irdma_sc_qp *qp, 83 struct irdma_sc_qp *qp = &rsrc->qp; local 89 qp->qp_uk.rq_wrid_array[wqe_idx] = (uintptr_t)buf; 90 wqe = qp->qp_uk.rq_base[wqe_idx].elem; 98 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) { 188 * @qp: puda qp for wqe 191 static __le64 *irdma_puda_get_next_send_wqe(struct irdma_qp_uk *qp, argument 196 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp 326 struct irdma_qp_uk *qp; local 425 irdma_puda_send(struct irdma_sc_qp *qp, struct irdma_puda_send_info *info) argument 568 struct irdma_sc_qp *qp = &rsrc->qp; local 605 irdma_puda_qp_wqe(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument 646 struct irdma_sc_qp *qp = &rsrc->qp; local 1120 irdma_ilq_putback_rcvbuf(struct irdma_sc_qp *qp, struct irdma_puda_buf *buf, u32 wqe_idx) argument 1532 irdma_ieq_process_fpdus(struct irdma_sc_qp *qp, struct irdma_puda_rsrc *ieq) argument 1575 irdma_ieq_create_ah(struct irdma_sc_qp *qp, struct irdma_puda_buf *buf) argument 1592 irdma_ieq_handle_exception(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp, struct irdma_puda_buf *buf) argument 1683 struct irdma_sc_qp *qp = NULL; local 1720 irdma_ieq_cleanup_qp(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp) argument [all...] |
H A D | osdep.h | 41 void irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp); 50 void irdma_send_ieq_ack(struct irdma_sc_qp *qp); 69 void irdma_term_modify_qp(struct irdma_sc_qp *qp, u8 next_state, u8 term, 71 void irdma_terminate_done(struct irdma_sc_qp *qp, int timeout_occurred); 72 void irdma_terminate_start_timer(struct irdma_sc_qp *qp); 73 void irdma_terminate_del_timer(struct irdma_sc_qp *qp);
|
H A D | protos.h | 20 int irdma_sc_mr_fast_register(struct irdma_sc_qp *qp, 59 int irdma_cqp_qp_suspend_resume(struct irdma_sc_qp *qp, u8 cmd); 60 void irdma_qp_add_qos(struct irdma_sc_qp *qp); 61 void irdma_qp_rem_qos(struct irdma_sc_qp *qp); 63 struct irdma_sc_qp *qp); 66 void irdma_terminate_send_fin(struct irdma_sc_qp *qp); 68 void irdma_terminate_connection(struct irdma_sc_qp *qp, 71 void irdma_terminate_received(struct irdma_sc_qp *qp,
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | gsi.c | 82 wr->wc.qp = &mqp->ibqp; 219 static int modify_to_rts(struct mlx5_ib_gsi_qp *gsi, struct ib_qp *qp, argument 222 struct mlx5_ib_dev *dev = to_mdev(qp->device); 232 ret = ib_modify_qp(qp, &attr, mask); 235 qp->qp_num, ret); 240 ret = ib_modify_qp(qp, &attr, IB_QP_STATE); 243 qp->qp_num, ret); 249 ret = ib_modify_qp(qp, &attr, IB_QP_STATE | IB_QP_SQ_PSN); 252 qp->qp_num, ret); 265 struct ib_qp *qp; local 320 mlx5_ib_gsi_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr, int attr_mask) argument 345 mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument 429 mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 474 mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/include/rdma/ |
H A D | rdma_vt.h | 35 struct rvt_qp __rcu *qp[2]; member in struct:rvt_ibport 213 bool (*schedule_send)(struct rvt_qp *qp); 214 bool (*schedule_send_no_lock)(struct rvt_qp *qp); 223 int (*setup_wqe)(struct rvt_qp *qp, struct rvt_swqe *wqe, 230 void (*do_send)(struct rvt_qp *qp); 245 void * (*qp_priv_alloc)(struct rvt_dev_info *rdi, struct rvt_qp *qp); 249 * called after all qp fields have been initialized in rdmavt. 251 int (*qp_priv_init)(struct rvt_dev_info *rdi, struct rvt_qp *qp, 255 * Free the driver's private qp structure. 257 void (*qp_priv_free)(struct rvt_dev_info *rdi, struct rvt_qp *qp); [all...] |
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | Makefile | 4 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
|
/linux-master/drivers/infiniband/core/ |
H A D | iwcm.h | 50 struct ib_qp *qp; member in struct:iwcm_id_private
|
H A D | rw.c | 91 static int rdma_rw_init_one_mr(struct ib_qp *qp, u32 port_num, argument 95 u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, 96 qp->integrity_en); 100 reg->mr = ib_mr_pool_get(qp, &qp->rdma_mrs); 108 ib_mr_pool_put(qp, &qp->rdma_mrs, reg->mr); 115 if (rdma_protocol_iwarp(qp->device, port_num)) 124 static int rdma_rw_init_mr_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, argument 129 u32 pages_per_mr = rdma_rw_fr_page_list_len(qp 195 rdma_rw_init_map_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 sg_cnt, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 251 rdma_rw_init_single_wr(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 292 rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, u32 sg_offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 364 rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, struct ib_sig_attrs *sig_attrs, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 497 rdma_rw_ctx_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) argument 555 rdma_rw_ctx_post(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) argument 574 rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, enum dma_data_direction dir) argument 613 rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, enum dma_data_direction dir) argument 687 rdma_rw_init_mrs(struct ib_qp *qp, struct ib_qp_init_attr *attr) argument 730 rdma_rw_cleanup_mrs(struct ib_qp *qp) argument [all...] |
H A D | iwcm.c | 269 static int iwcm_modify_qp_err(struct ib_qp *qp) argument 273 if (!qp) 277 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); 284 static int iwcm_modify_qp_sqd(struct ib_qp *qp) argument 288 BUG_ON(qp == NULL); 290 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); 310 struct ib_qp *qp = NULL; local 323 if (cm_id_priv->qp) 324 qp = cm_id_priv->qp; 375 struct ib_qp *qp; local 644 struct ib_qp *qp; local 701 struct ib_qp *qp = NULL; local 883 struct ib_qp *qp = NULL; local 950 struct ib_qp *qp; local [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_qp_tx.c | 693 * @qp: QP from which to transmit 700 static void siw_prepare_fpdu(struct siw_qp *qp, struct siw_wqe *wqe) argument 702 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; 788 static int siw_qp_sq_proc_tx(struct siw_qp *qp, struct siw_wqe *wqe) argument 790 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; 791 struct socket *s = qp->attrs.sk; 792 int rv = 0, burst_len = qp->tx_ctx.burst; 813 rv = siw_check_sgl_tx(qp->pd, wqe, 0); 827 if (!rdma_is_kernel_res(&qp->base_qp.res)) { 844 siw_prepare_fpdu(qp, wq 973 siw_qp_sq_proc_local(struct siw_qp *qp, struct siw_wqe *wqe) argument 1018 siw_qp_sq_process(struct siw_qp *qp) argument 1167 siw_sq_resume(struct siw_qp *qp) argument 1243 struct siw_qp *qp; local 1280 siw_sq_start(struct siw_qp *qp) argument [all...] |
H A D | siw_cm.c | 68 static void siw_qp_socket_assoc(struct siw_cep *cep, struct siw_qp *qp) argument 75 qp->attrs.sk = s; 105 struct siw_qp *qp = NULL; local 117 qp = sk_to_qp(sk); 120 rd_desc.arg.data = qp; 130 if (!qp->rx_stream.rx_suspend) 134 if (qp) 135 siw_qp_socket_assoc(cep, qp); 356 cep->qp ? qp_id(cep->qp) 387 siw_qp_cm_drop(struct siw_qp *qp, int schedule) argument 731 struct siw_qp *qp = cep->qp; local 1166 struct siw_qp *qp = cep->qp; local 1355 struct siw_qp *qp; local 1555 struct siw_qp *qp; local [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | verbs.h | 227 struct rvt_qp __rcu *qp[2]; member in struct:hfi1_ibport 361 int hfi1_verbs_send(struct rvt_qp *qp, struct hfi1_pkt_state *ps); 372 struct rvt_qp *qp); 376 void hfi1_rc_verbs_aborted(struct rvt_qp *qp, struct hfi1_opa_header *opah); 377 void hfi1_rc_send_complete(struct rvt_qp *qp, struct hfi1_opa_header *opah); 383 void hfi1_migrate_qp(struct rvt_qp *qp); 385 int hfi1_check_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr, 388 void hfi1_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr, 390 void hfi1_restart_rc(struct rvt_qp *qp, u32 psn, int wait); 391 int hfi1_setup_wqe(struct rvt_qp *qp, struc 447 hfi1_trdma_send_complete(struct rvt_qp *qp, struct rvt_swqe *wqe, enum ib_wc_status status) argument [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.h | 38 #include <linux/mlx5/qp.h> 83 } qp; member in struct:mlx5_fpga_conn
|
/linux-master/drivers/crypto/hisilicon/zip/ |
H A D | zip_crypto.c | 62 struct hisi_qp *qp; member in struct:hisi_zip_qp_ctx 124 dev_dbg(&qp_ctx->qp->qm->pdev->dev, "req cache is full!\n"); 217 struct hisi_qp *qp = qp_ctx->qp; local 218 struct device *dev = &qp->qm->pdev->dev; 243 hisi_zip_fill_sqe(qp_ctx->ctx, &zip_sqe, qp->req_type, req); 247 ret = hisi_qp_send(qp, &zip_sqe); 279 static void hisi_zip_acomp_cb(struct hisi_qp *qp, void *data) argument 281 struct hisi_zip_qp_ctx *qp_ctx = qp->qp_ctx; 285 struct device *dev = &qp 357 hisi_zip_start_qp(struct hisi_qp *qp, struct hisi_zip_qp_ctx *qp_ctx, int alg_type, int req_type) argument [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 39 static int create_qp_cmd(struct erdma_ucontext *uctx, struct erdma_qp *qp) argument 41 struct erdma_dev *dev = to_edev(qp->ibqp.device); 42 struct erdma_pd *pd = to_epd(qp->ibqp.pd); 52 ilog2(qp->attrs.sq_size)) | 53 FIELD_PREP(ERDMA_CMD_CREATE_QP_QPN_MASK, QP_ID(qp)); 55 ilog2(qp->attrs.rq_size)) | 58 if (rdma_is_kernel_res(&qp->ibqp.res)) { 64 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->scq->cqn); 68 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->rcq->cqn); 77 req.rq_buf_addr = qp 432 struct erdma_qp *qp = local 474 free_kernel_qp(struct erdma_qp *qp) argument 502 init_kernel_qp(struct erdma_dev *dev, struct erdma_qp *qp, struct ib_qp_init_attr *attrs) argument 878 init_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx, u64 va, u32 len, u64 dbrec_va) argument 924 free_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx) argument 934 struct erdma_qp *qp = to_eqp(ibqp); local 1268 struct erdma_qp *qp = to_eqp(ibqp); local 1524 struct erdma_qp *qp = to_eqp(ibqp); local 1550 struct erdma_qp *qp; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 40 #include <linux/mlx4/qp.h> 49 void mlx4_put_qp(struct mlx4_qp *qp) argument 51 if (refcount_dec_and_test(&qp->refcount)) 52 complete(&qp->free); 59 struct mlx4_qp *qp; local 63 qp = __mlx4_qp_lookup(dev, qpn); 64 if (qp) 65 refcount_inc(&qp->refcount); 69 if (!qp) { 75 qp 79 is_master_qp0(struct mlx4_dev *dev, struct mlx4_qp *qp, int *real_qp0, int *proxy_qp0) argument 92 __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp, int native) argument 213 mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp) argument 399 struct mlx4_qp *qp; local 409 mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) argument 518 mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp) argument 529 mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) argument 898 mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, struct mlx4_qp_context *context) argument 919 mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_qp_context *context, struct mlx4_qp *qp, enum mlx4_qp_state *qp_state) argument 955 struct mlx4_qp qp; local [all...] |