Searched refs:qp (Results 76 - 100 of 380) sorted by relevance

1234567891011>>

/linux-master/drivers/interconnect/qcom/
H A Dosm-l3.c128 struct qcom_osm_l3_icc_provider *qp; local
136 qp = to_osm_l3_provider(provider);
141 for (index = 0; index < qp->max_state - 1; index++) {
142 if (qp->lut_tables[index] >= rate)
146 writel_relaxed(index, qp->base + qp->reg_perf_state);
153 struct qcom_osm_l3_icc_provider *qp = platform_get_drvdata(pdev); local
155 icc_provider_deregister(&qp->provider);
156 icc_nodes_remove(&qp->provider);
163 struct qcom_osm_l3_icc_provider *qp; local
[all...]
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_net.c23 static struct dst_entry *rxe_find_route4(struct rxe_qp *qp, argument
39 rxe_dbg_qp(qp, "no route to %pI4\n", &daddr->s_addr);
47 static struct dst_entry *rxe_find_route6(struct rxe_qp *qp, argument
65 rxe_dbg_qp(qp, "no route to %pI6\n", daddr);
70 rxe_dbg_qp(qp, "no route to %pI6\n", daddr);
82 static struct dst_entry *rxe_find_route6(struct rxe_qp *qp, argument
93 struct rxe_qp *qp,
98 if (qp_type(qp) == IB_QPT_RC)
99 dst = sk_dst_get(qp->sk->sk);
101 if (!dst || !dst_check(dst, qp
92 rxe_find_route(struct net_device *ndev, struct rxe_qp *qp, struct rxe_av *av) argument
282 struct rxe_qp *qp = pkt->qp; local
308 struct rxe_qp *qp = pkt->qp; local
351 struct rxe_qp *qp; local
428 rxe_xmit_packet(struct rxe_qp *qp, struct rxe_pkt_info *pkt, struct sk_buff *skb) argument
[all...]
H A Drxe_verbs.c526 /* qp */
532 struct rxe_qp *qp = to_rqp(ibqp); local
549 qp->is_user = true;
552 qp->is_user = false;
567 err = rxe_add_to_pool(&rxe->qp_pool, qp);
569 rxe_dbg_dev(rxe, "unable to create qp, err = %d\n", err);
573 err = rxe_qp_from_init(rxe, qp, pd, init, uresp, ibqp->pd, udata);
575 rxe_dbg_qp(qp, "create qp failed, err = %d\n", err);
579 rxe_finalize(qp);
595 struct rxe_qp *qp = to_rqp(ibqp); local
632 struct rxe_qp *qp = to_rqp(ibqp); local
642 struct rxe_qp *qp = to_rqp(ibqp); local
665 validate_send_wr(struct rxe_qp *qp, const struct ib_send_wr *ibwr, unsigned int *maskp, unsigned int *lengthp) argument
726 init_send_wr(struct rxe_qp *qp, struct rxe_send_wr *wr, const struct ib_send_wr *ibwr) argument
820 init_send_wqe(struct rxe_qp *qp, const struct ib_send_wr *ibwr, unsigned int mask, unsigned int length, struct rxe_send_wqe *wqe) argument
858 post_one_send(struct rxe_qp *qp, const struct ib_send_wr *ibwr) argument
885 rxe_post_send_kernel(struct rxe_qp *qp, const struct ib_send_wr *ibwr, const struct ib_send_wr **bad_wr) argument
916 struct rxe_qp *qp = to_rqp(ibqp); local
1006 struct rxe_qp *qp = to_rqp(ibqp); local
[all...]
H A Drxe_mcast.c11 * struct rxe_mca ('mca'). An mcg is allocated each time a qp is
15 * qp is attached to the same mgid. It is cleaned up when the last qp
16 * is detached from the mcg. Each time a qp is attached to an mcg an
17 * mca is created. It holds a pointer to the qp and is added to a list
18 * of qp's that are attached to the mcg. The qp_list is used to replicate
254 * no qp's are attached to mcg
271 * Context: no qp's are attached to mcg
285 * @qp: qp objec
294 __rxe_init_mca(struct rxe_qp *qp, struct rxe_mcg *mcg, struct rxe_mca *mca) argument
331 rxe_attach_mcg(struct rxe_mcg *mcg, struct rxe_qp *qp) argument
396 rxe_detach_mcg(struct rxe_mcg *mcg, struct rxe_qp *qp) argument
437 struct rxe_qp *qp = to_rqp(ibqp); local
467 struct rxe_qp *qp = to_rqp(ibqp); local
[all...]
H A Drxe_task.c28 * a qp reference to cover the gap from now until the task finishes.
38 WARN_ON(rxe_read(task->qp) <= 0);
41 rxe_get(task->qp);
111 WARN_ON(rxe_read(task->qp) <= 0);
115 rxe_put(task->qp);
127 ret = task->func(task->qp);
159 rxe_dbg_qp(task->qp, "unexpected task state = %d\n",
169 task->qp,
181 rxe_put(task->qp);
190 int rxe_init_task(struct rxe_task *task, struct rxe_qp *qp, argument
[all...]
/linux-master/drivers/infiniband/sw/rdmavt/
H A Dtrace_tx.h49 TP_PROTO(struct rvt_qp *qp, struct rvt_swqe *wqe, int wr_num_sge),
50 TP_ARGS(qp, wqe, wr_num_sge),
52 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device))
72 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device));
75 __entry->qpn = qp->ibqp.qp_num;
76 __entry->qpt = qp->ibqp.qp_type;
81 __entry->size = qp->s_size;
82 __entry->avail = qp->s_avail;
83 __entry->head = qp->s_head;
84 __entry->last = qp
[all...]
H A Dmcast.c31 * @qp: the QP to link
33 static struct rvt_mcast_qp *rvt_mcast_qp_alloc(struct rvt_qp *qp) argument
41 mqp->qp = qp;
42 rvt_get_qp(qp);
50 struct rvt_qp *qp = mqp->qp; local
53 rvt_put_qp(qp);
186 if (p->qp == mqp->qp) {
240 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); local
303 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); local
[all...]
H A DMakefile11 rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dib_verbs.c429 struct bnxt_re_qp *qp = container_of(qplib_qp, struct bnxt_re_qp, local
431 struct ib_pd *ib_pd = qp->ib_qp.pd;
442 ibdev_dbg(&qp->rdev->ibdev,
444 wqe.bind.r_key, qp->qplib_qp.id, pd);
445 rc = bnxt_qplib_post_send(&qp->qplib_qp, &wqe);
447 ibdev_err(&qp->rdev->ibdev, "Failed to bind fence-WQE\n");
450 bnxt_qplib_post_send_db(&qp->qplib_qp);
819 unsigned long bnxt_re_lock_cqs(struct bnxt_re_qp *qp)
820 __acquires(&qp->scq->cq_lock) __acquires(&qp
844 bnxt_re_destroy_gsi_sqp(struct bnxt_re_qp *qp) argument
891 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
980 bnxt_re_setup_swqe_size(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr) argument
1019 bnxt_re_init_user_qp(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd, struct bnxt_re_qp *qp, struct ib_udata *udata) argument
1137 struct bnxt_re_qp *qp; local
1198 bnxt_re_init_rq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument
1239 bnxt_re_adjust_gsi_rq_attr(struct bnxt_re_qp *qp) argument
1257 bnxt_re_init_sq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument
1303 bnxt_re_adjust_gsi_sq_attr(struct bnxt_re_qp *qp, struct ib_qp_init_attr *init_attr, struct bnxt_re_ucontext *uctx) argument
1350 bnxt_re_init_qp_attr(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd, struct ib_qp_init_attr *init_attr, struct ib_udata *udata) argument
1426 bnxt_re_create_shadow_gsi(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd) argument
1471 bnxt_re_create_gsi_qp(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd, struct ib_qp_init_attr *init_attr) argument
1526 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
1907 struct bnxt_re_qp *qp = rdev->gsi_ctx.gsi_sqp; local
1938 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
2183 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
2241 bnxt_re_build_qp1_send_v2(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument
2415 bnxt_re_build_qp1_shadow_qp_recv(struct bnxt_re_qp *qp, const struct ib_recv_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument
2458 is_ud_qp(struct bnxt_re_qp *qp) argument
2464 bnxt_re_build_send_wqe(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument
2657 bnxt_ud_qp_hw_stall_workaround(struct bnxt_re_qp *qp) argument
2673 bnxt_re_post_send_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_send_wr *wr) argument
2735 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
2833 bnxt_re_post_recv_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_recv_wr *wr) argument
2870 struct bnxt_re_qp *qp = container_of(ib_qp, struct bnxt_re_qp, ib_qp); local
3609 bnxt_re_process_res_ud_wc(struct bnxt_re_qp *qp, struct ib_wc *wc, struct bnxt_qplib_cqe *cqe) argument
3644 send_phantom_wqe(struct bnxt_re_qp *qp) argument
3669 struct bnxt_re_qp *qp, *sh_qp; local
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Dpuda.c14 static void irdma_ilq_putback_rcvbuf(struct irdma_sc_qp *qp,
83 struct irdma_sc_qp *qp = &rsrc->qp; local
89 qp->qp_uk.rq_wrid_array[wqe_idx] = (uintptr_t)buf;
90 wqe = qp->qp_uk.rq_base[wqe_idx].elem;
98 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) {
188 * @qp: puda qp for wqe
191 static __le64 *irdma_puda_get_next_send_wqe(struct irdma_qp_uk *qp, argument
196 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp
326 struct irdma_qp_uk *qp; local
425 irdma_puda_send(struct irdma_sc_qp *qp, struct irdma_puda_send_info *info) argument
568 struct irdma_sc_qp *qp = &rsrc->qp; local
605 irdma_puda_qp_wqe(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument
646 struct irdma_sc_qp *qp = &rsrc->qp; local
1120 irdma_ilq_putback_rcvbuf(struct irdma_sc_qp *qp, struct irdma_puda_buf *buf, u32 wqe_idx) argument
1532 irdma_ieq_process_fpdus(struct irdma_sc_qp *qp, struct irdma_puda_rsrc *ieq) argument
1575 irdma_ieq_create_ah(struct irdma_sc_qp *qp, struct irdma_puda_buf *buf) argument
1592 irdma_ieq_handle_exception(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp, struct irdma_puda_buf *buf) argument
1683 struct irdma_sc_qp *qp = NULL; local
1720 irdma_ieq_cleanup_qp(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp) argument
[all...]
H A Dosdep.h41 void irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp);
50 void irdma_send_ieq_ack(struct irdma_sc_qp *qp);
69 void irdma_term_modify_qp(struct irdma_sc_qp *qp, u8 next_state, u8 term,
71 void irdma_terminate_done(struct irdma_sc_qp *qp, int timeout_occurred);
72 void irdma_terminate_start_timer(struct irdma_sc_qp *qp);
73 void irdma_terminate_del_timer(struct irdma_sc_qp *qp);
H A Dprotos.h20 int irdma_sc_mr_fast_register(struct irdma_sc_qp *qp,
59 int irdma_cqp_qp_suspend_resume(struct irdma_sc_qp *qp, u8 cmd);
60 void irdma_qp_add_qos(struct irdma_sc_qp *qp);
61 void irdma_qp_rem_qos(struct irdma_sc_qp *qp);
63 struct irdma_sc_qp *qp);
66 void irdma_terminate_send_fin(struct irdma_sc_qp *qp);
68 void irdma_terminate_connection(struct irdma_sc_qp *qp,
71 void irdma_terminate_received(struct irdma_sc_qp *qp,
/linux-master/drivers/infiniband/hw/mlx5/
H A Dgsi.c82 wr->wc.qp = &mqp->ibqp;
219 static int modify_to_rts(struct mlx5_ib_gsi_qp *gsi, struct ib_qp *qp, argument
222 struct mlx5_ib_dev *dev = to_mdev(qp->device);
232 ret = ib_modify_qp(qp, &attr, mask);
235 qp->qp_num, ret);
240 ret = ib_modify_qp(qp, &attr, IB_QP_STATE);
243 qp->qp_num, ret);
249 ret = ib_modify_qp(qp, &attr, IB_QP_STATE | IB_QP_SQ_PSN);
252 qp->qp_num, ret);
265 struct ib_qp *qp; local
320 mlx5_ib_gsi_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr, int attr_mask) argument
345 mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument
429 mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument
474 mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument
[all...]
/linux-master/include/rdma/
H A Drdma_vt.h35 struct rvt_qp __rcu *qp[2]; member in struct:rvt_ibport
213 bool (*schedule_send)(struct rvt_qp *qp);
214 bool (*schedule_send_no_lock)(struct rvt_qp *qp);
223 int (*setup_wqe)(struct rvt_qp *qp, struct rvt_swqe *wqe,
230 void (*do_send)(struct rvt_qp *qp);
245 void * (*qp_priv_alloc)(struct rvt_dev_info *rdi, struct rvt_qp *qp);
249 * called after all qp fields have been initialized in rdmavt.
251 int (*qp_priv_init)(struct rvt_dev_info *rdi, struct rvt_qp *qp,
255 * Free the driver's private qp structure.
257 void (*qp_priv_free)(struct rvt_dev_info *rdi, struct rvt_qp *qp);
[all...]
/linux-master/drivers/infiniband/hw/mlx4/
H A DMakefile4 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
/linux-master/drivers/infiniband/core/
H A Diwcm.h50 struct ib_qp *qp; member in struct:iwcm_id_private
H A Drw.c91 static int rdma_rw_init_one_mr(struct ib_qp *qp, u32 port_num, argument
95 u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
96 qp->integrity_en);
100 reg->mr = ib_mr_pool_get(qp, &qp->rdma_mrs);
108 ib_mr_pool_put(qp, &qp->rdma_mrs, reg->mr);
115 if (rdma_protocol_iwarp(qp->device, port_num))
124 static int rdma_rw_init_mr_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, argument
129 u32 pages_per_mr = rdma_rw_fr_page_list_len(qp
195 rdma_rw_init_map_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 sg_cnt, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument
251 rdma_rw_init_single_wr(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument
292 rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, u32 sg_offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument
364 rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, struct ib_sig_attrs *sig_attrs, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument
497 rdma_rw_ctx_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) argument
555 rdma_rw_ctx_post(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) argument
574 rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, enum dma_data_direction dir) argument
613 rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, enum dma_data_direction dir) argument
687 rdma_rw_init_mrs(struct ib_qp *qp, struct ib_qp_init_attr *attr) argument
730 rdma_rw_cleanup_mrs(struct ib_qp *qp) argument
[all...]
H A Diwcm.c269 static int iwcm_modify_qp_err(struct ib_qp *qp) argument
273 if (!qp)
277 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE);
284 static int iwcm_modify_qp_sqd(struct ib_qp *qp) argument
288 BUG_ON(qp == NULL);
290 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE);
310 struct ib_qp *qp = NULL; local
323 if (cm_id_priv->qp)
324 qp = cm_id_priv->qp;
375 struct ib_qp *qp; local
644 struct ib_qp *qp; local
701 struct ib_qp *qp = NULL; local
883 struct ib_qp *qp = NULL; local
950 struct ib_qp *qp; local
[all...]
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_qp_tx.c693 * @qp: QP from which to transmit
700 static void siw_prepare_fpdu(struct siw_qp *qp, struct siw_wqe *wqe) argument
702 struct siw_iwarp_tx *c_tx = &qp->tx_ctx;
788 static int siw_qp_sq_proc_tx(struct siw_qp *qp, struct siw_wqe *wqe) argument
790 struct siw_iwarp_tx *c_tx = &qp->tx_ctx;
791 struct socket *s = qp->attrs.sk;
792 int rv = 0, burst_len = qp->tx_ctx.burst;
813 rv = siw_check_sgl_tx(qp->pd, wqe, 0);
827 if (!rdma_is_kernel_res(&qp->base_qp.res)) {
844 siw_prepare_fpdu(qp, wq
973 siw_qp_sq_proc_local(struct siw_qp *qp, struct siw_wqe *wqe) argument
1018 siw_qp_sq_process(struct siw_qp *qp) argument
1167 siw_sq_resume(struct siw_qp *qp) argument
1243 struct siw_qp *qp; local
1280 siw_sq_start(struct siw_qp *qp) argument
[all...]
H A Dsiw_cm.c68 static void siw_qp_socket_assoc(struct siw_cep *cep, struct siw_qp *qp) argument
75 qp->attrs.sk = s;
105 struct siw_qp *qp = NULL; local
117 qp = sk_to_qp(sk);
120 rd_desc.arg.data = qp;
130 if (!qp->rx_stream.rx_suspend)
134 if (qp)
135 siw_qp_socket_assoc(cep, qp);
356 cep->qp ? qp_id(cep->qp)
387 siw_qp_cm_drop(struct siw_qp *qp, int schedule) argument
731 struct siw_qp *qp = cep->qp; local
1166 struct siw_qp *qp = cep->qp; local
1355 struct siw_qp *qp; local
1555 struct siw_qp *qp; local
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dverbs.h227 struct rvt_qp __rcu *qp[2]; member in struct:hfi1_ibport
361 int hfi1_verbs_send(struct rvt_qp *qp, struct hfi1_pkt_state *ps);
372 struct rvt_qp *qp);
376 void hfi1_rc_verbs_aborted(struct rvt_qp *qp, struct hfi1_opa_header *opah);
377 void hfi1_rc_send_complete(struct rvt_qp *qp, struct hfi1_opa_header *opah);
383 void hfi1_migrate_qp(struct rvt_qp *qp);
385 int hfi1_check_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr,
388 void hfi1_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr,
390 void hfi1_restart_rc(struct rvt_qp *qp, u32 psn, int wait);
391 int hfi1_setup_wqe(struct rvt_qp *qp, struc
447 hfi1_trdma_send_complete(struct rvt_qp *qp, struct rvt_swqe *wqe, enum ib_wc_status status) argument
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.h38 #include <linux/mlx5/qp.h>
83 } qp; member in struct:mlx5_fpga_conn
/linux-master/drivers/crypto/hisilicon/zip/
H A Dzip_crypto.c62 struct hisi_qp *qp; member in struct:hisi_zip_qp_ctx
124 dev_dbg(&qp_ctx->qp->qm->pdev->dev, "req cache is full!\n");
217 struct hisi_qp *qp = qp_ctx->qp; local
218 struct device *dev = &qp->qm->pdev->dev;
243 hisi_zip_fill_sqe(qp_ctx->ctx, &zip_sqe, qp->req_type, req);
247 ret = hisi_qp_send(qp, &zip_sqe);
279 static void hisi_zip_acomp_cb(struct hisi_qp *qp, void *data) argument
281 struct hisi_zip_qp_ctx *qp_ctx = qp->qp_ctx;
285 struct device *dev = &qp
357 hisi_zip_start_qp(struct hisi_qp *qp, struct hisi_zip_qp_ctx *qp_ctx, int alg_type, int req_type) argument
[all...]
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_verbs.c39 static int create_qp_cmd(struct erdma_ucontext *uctx, struct erdma_qp *qp) argument
41 struct erdma_dev *dev = to_edev(qp->ibqp.device);
42 struct erdma_pd *pd = to_epd(qp->ibqp.pd);
52 ilog2(qp->attrs.sq_size)) |
53 FIELD_PREP(ERDMA_CMD_CREATE_QP_QPN_MASK, QP_ID(qp));
55 ilog2(qp->attrs.rq_size)) |
58 if (rdma_is_kernel_res(&qp->ibqp.res)) {
64 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->scq->cqn);
68 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->rcq->cqn);
77 req.rq_buf_addr = qp
432 struct erdma_qp *qp = local
474 free_kernel_qp(struct erdma_qp *qp) argument
502 init_kernel_qp(struct erdma_dev *dev, struct erdma_qp *qp, struct ib_qp_init_attr *attrs) argument
878 init_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx, u64 va, u32 len, u64 dbrec_va) argument
924 free_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx) argument
934 struct erdma_qp *qp = to_eqp(ibqp); local
1268 struct erdma_qp *qp = to_eqp(ibqp); local
1524 struct erdma_qp *qp = to_eqp(ibqp); local
1550 struct erdma_qp *qp; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Dqp.c40 #include <linux/mlx4/qp.h>
49 void mlx4_put_qp(struct mlx4_qp *qp) argument
51 if (refcount_dec_and_test(&qp->refcount))
52 complete(&qp->free);
59 struct mlx4_qp *qp; local
63 qp = __mlx4_qp_lookup(dev, qpn);
64 if (qp)
65 refcount_inc(&qp->refcount);
69 if (!qp) {
75 qp
79 is_master_qp0(struct mlx4_dev *dev, struct mlx4_qp *qp, int *real_qp0, int *proxy_qp0) argument
92 __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp, int native) argument
213 mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp) argument
399 struct mlx4_qp *qp; local
409 mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) argument
518 mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp) argument
529 mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) argument
898 mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, struct mlx4_qp_context *context) argument
919 mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_qp_context *context, struct mlx4_qp *qp, enum mlx4_qp_state *qp_state) argument
955 struct mlx4_qp qp; local
[all...]

Completed in 241 milliseconds

1234567891011>>