Searched refs:qp (Results 101 - 125 of 380) sorted by relevance

1234567891011>>

/linux-master/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.h46 int usnic_ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
53 int usnic_ib_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *init_attr,
55 int usnic_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
/linux-master/drivers/infiniband/hw/qib/
H A Dqib_driver.c301 struct rvt_qp *qp = NULL; local
346 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num);
347 if (!qp) {
356 spin_lock(&qp->r_lock);
359 if (!(ib_rvt_state_ops[qp->state] &
365 switch (qp->ibqp.qp_type) {
371 qp,
379 diff = qib_cmp24(psn, qp->r_psn);
380 if (!qp->r_nak_state && diff >= 0) {
382 qp
448 struct rvt_qp *qp, *nqp; local
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dverbs.c21 #include "qp.h"
101 static int pio_wait(struct rvt_qp *qp,
276 if (!(ib_rvt_state_ops[packet->qp->state] & RVT_PROCESS_RECV_OK))
279 packet->qp->allowed_ops) ||
286 static u64 hfi1_fault_tx(struct rvt_qp *qp, u8 opcode, u64 pbc) argument
319 if (packet->qp->ibqp.qp_type != IB_QPT_RC ||
320 !(ib_rvt_state_ops[packet->qp->state] & RVT_PROCESS_RECV_OK))
360 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num);
361 if (!packet->qp)
363 spin_lock_irqsave(&packet->qp
559 struct rvt_qp *qp = NULL; local
591 struct rvt_qp *qp = tx->qp; local
609 hfi1_wait_kmem(struct rvt_qp *qp) argument
627 wait_kmem(struct hfi1_ibdev *dev, struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument
697 update_tx_opstats(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u32 plen) argument
797 hfi1_verbs_send_dma(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc) argument
881 pio_wait(struct rvt_qp *qp, struct send_context *sc, struct hfi1_pkt_state *ps, u32 flag) argument
930 struct rvt_qp *qp = (struct rvt_qp *)arg; local
937 hfi1_verbs_send_pio(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc) argument
1173 get_send_routine(struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument
1213 hfi1_verbs_send(struct rvt_qp *qp, struct hfi1_pkt_state *ps) argument
1917 struct rvt_qp *qp = packet->qp; local
[all...]
H A Dtrace_misc.h56 TP_PROTO(struct rvt_qp *qp, u8 opcode),
57 TP_ARGS(qp, opcode),
58 TP_STRUCT__entry(DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
62 TP_fast_assign(DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device));
63 __entry->qpn = qp->ibqp.qp_num;
/linux-master/drivers/infiniband/hw/mlx5/
H A Dqpc.c7 #include <linux/mlx5/qp.h>
10 #include "qp.h"
114 struct mlx5_core_qp *qp; local
146 qp = (struct mlx5_core_qp *)common;
147 qp->event(qp, event_type);
160 struct mlx5_core_qp *qp, int rsc_type)
165 qp->common.res = rsc_type;
168 qp->qpn | (rsc_type << MLX5_USER_INDEX_LEN),
169 qp);
159 create_resource_common(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, int rsc_type) argument
181 destroy_resource_common(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp) argument
199 struct mlx5_core_qp *qp = &dct->mqp; local
210 struct mlx5_core_qp *qp = &dct->mqp; local
232 mlx5_qpc_create_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, u32 *in, int inlen, u32 *out) argument
268 struct mlx5_core_qp *qp = &dct->mqp; local
306 mlx5_core_destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp) argument
479 mlx5_core_qp_modify(struct mlx5_ib_dev *dev, u16 opcode, u32 opt_param_mask, void *qpc, struct mlx5_core_qp *qp, u32 *ece) argument
523 mlx5_core_qp_query(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, u32 *out, int outlen, bool qpc_ext) argument
539 struct mlx5_core_qp *qp = &dct->mqp; local
[all...]
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_cm.c270 void erdma_qp_cm_drop(struct erdma_qp *qp) argument
272 struct erdma_cep *cep = qp->cep;
274 if (!qp->cep)
313 if (cep->qp) {
314 cep->qp = NULL;
315 erdma_qp_put(qp);
571 struct erdma_qp *qp = cep->qp; local
605 down_write(&qp->state_lock);
606 if (qp
858 struct erdma_qp *qp = cep->qp; local
1003 struct erdma_qp *qp; local
1131 struct erdma_qp *qp; local
[all...]
H A Derdma_cq.c116 struct erdma_qp *qp; local
135 qp = find_qp_by_qpn(dev, qpn);
136 if (!qp)
139 kern_qp = &qp->kern_qp;
147 depth = qp->attrs.sq_size;
148 wqe_hdr = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx,
149 qp->attrs.sq_size, SQEBB_SHIFT);
155 depth = qp->attrs.rq_size;
176 wc->qp = &qp
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Duser.h258 int irdma_uk_inline_rdma_write(struct irdma_qp_uk *qp,
260 int irdma_uk_inline_send(struct irdma_qp_uk *qp,
262 int irdma_uk_post_nop(struct irdma_qp_uk *qp, u64 wr_id, bool signaled,
264 int irdma_uk_post_receive(struct irdma_qp_uk *qp,
266 void irdma_uk_qp_post_wr(struct irdma_qp_uk *qp);
267 int irdma_uk_rdma_read(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
269 int irdma_uk_rdma_write(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
271 int irdma_uk_send(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
273 int irdma_uk_stag_local_invalidate(struct irdma_qp_uk *qp,
292 void irdma_uk_cq_set_resized_cnt(struct irdma_cq_uk *qp, u1
[all...]
H A Dctrl.c13 * irdma_get_qp_from_list - get next qp from a list
14 * @head: Listhead of qp's
15 * @qp: current qp
18 struct irdma_sc_qp *qp)
26 if (!qp) {
29 lastentry = &qp->list;
39 * irdma_sc_suspend_resume_qps - suspend/resume all qp's on VSI
45 struct irdma_sc_qp *qp = NULL; local
50 qp
17 irdma_get_qp_from_list(struct list_head *head, struct irdma_sc_qp *qp) argument
123 irdma_qp_rem_qos(struct irdma_sc_qp *qp) argument
143 irdma_qp_add_qos(struct irdma_sc_qp *qp) argument
375 irdma_sc_qp_init(struct irdma_sc_qp *qp, struct irdma_qp_init_info *info) argument
448 irdma_sc_qp_create(struct irdma_sc_qp *qp, struct irdma_create_qp_info *info, u64 scratch, bool post_sq) argument
499 irdma_sc_qp_modify(struct irdma_sc_qp *qp, struct irdma_modify_qp_info *info, u64 scratch, bool post_sq) argument
569 irdma_sc_qp_destroy(struct irdma_sc_qp *qp, u64 scratch, bool remove_hash_idx, bool ignore_mw_bnd, bool post_sq) argument
637 irdma_sc_qp_setctx_roce(struct irdma_sc_qp *qp, __le64 *qp_ctx, struct irdma_qp_host_ctx_info *info) argument
869 irdma_sc_qp_setctx(struct irdma_sc_qp *qp, __le64 *qp_ctx, struct irdma_qp_host_ctx_info *info) argument
1291 irdma_sc_mr_fast_register(struct irdma_sc_qp *qp, struct irdma_fast_reg_stag_info *info, bool post_sq) argument
1364 irdma_sc_gen_rts_ae(struct irdma_sc_qp *qp) argument
1400 irdma_sc_send_lsmm(struct irdma_sc_qp *qp, void *lsmm_buf, u32 size, irdma_stag stag) argument
1443 irdma_sc_send_rtt(struct irdma_sc_qp *qp, bool read) argument
1530 irdma_bld_termhdr_ctrl(struct irdma_sc_qp *qp, struct irdma_terminate_hdr *hdr, enum irdma_flush_opcode opcode, u8 layer_etype, u8 err) argument
1581 irdma_bld_terminate_hdr(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument
1759 irdma_terminate_send_fin(struct irdma_sc_qp *qp) argument
1770 irdma_terminate_connection(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument
1790 irdma_terminate_received(struct irdma_sc_qp *qp, struct irdma_aeqe_info *info) argument
2215 irdma_sc_qp_flush_wqes(struct irdma_sc_qp *qp, struct irdma_qp_flush_info *info, u64 scratch, bool post_sq) argument
2289 irdma_sc_gen_ae(struct irdma_sc_qp *qp, struct irdma_gen_ae_info *info, u64 scratch, bool post_sq) argument
2408 irdma_sc_suspend_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) argument
2438 irdma_sc_resume_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) argument
[all...]
H A Dpuda.h30 struct irdma_qp_uk *qp; member in struct:irdma_puda_cmpl_info
101 struct irdma_sc_qp qp; member in struct:irdma_puda_rsrc
154 int irdma_puda_send(struct irdma_sc_qp *qp, struct irdma_puda_send_info *info);
168 void irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp);
171 int irdma_cqp_qp_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp);
173 int irdma_cqp_qp_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp);
175 void irdma_puda_ieq_get_ah_info(struct irdma_sc_qp *qp,
182 void irdma_ieq_process_fpdus(struct irdma_sc_qp *qp,
184 void irdma_ieq_cleanup_qp(struct irdma_puda_rsrc *ieq, struct irdma_sc_qp *qp);
/linux-master/drivers/infiniband/hw/cxgb4/
H A Ddevice.c245 static int dump_qp(unsigned long id, struct c4iw_qp *qp, argument
250 if (id != qp->wq.sq.qid)
257 if (qp->ep) {
258 struct c4iw_ep *ep = qp->ep;
268 "rc qp sq id %u %s id %u state %u "
271 qp->wq.sq.qid, qp->srq ? "srq" : "rq",
272 qp->srq ? qp->srq->idx : qp
328 struct c4iw_qp *qp; local
1270 struct c4iw_qp *qp; local
1285 resume_rc_qp(struct c4iw_qp *qp) argument
1298 struct c4iw_qp *qp; local
1318 struct c4iw_qp *qp; local
1374 struct c4iw_qp *qp = qp_list->qps[idx]; local
1417 struct c4iw_qp *qp; local
[all...]
H A DMakefile7 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o \
/linux-master/drivers/infiniband/hw/mana/
H A Dqp.c95 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local
120 "Failed copy from udata for create rss-qp, err %d\n",
158 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n",
174 qp->port = port;
237 "Failed to copy to udata create rss-qp, %d\n",
267 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local
292 "Failed to copy from udata create qp-raw, %d\n", err);
313 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n",
324 qp->port = port;
329 err = mana_ib_create_queue(mdev, ucmd.sq_buf_addr, ucmd.sq_buf_size, &qp
428 mana_ib_destroy_qp_rss(struct mana_ib_qp *qp, struct ib_rwq_ind_table *ind_tbl, struct ib_udata *udata) argument
454 mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata) argument
478 struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); local
[all...]
/linux-master/lib/crypto/mpi/
H A Dmpi-div.c86 mpi_ptr_t qp, rp; local
133 qp = quot->d;
134 rlimb = mpihelp_divmod_1(qp, np, nsize, dp[0]);
135 qsize -= qp[qsize - 1] == 0;
149 qp = quot->d;
153 if (qp == np) { /* Copy NP object to temporary space. */
155 MPN_COPY(np, qp, nsize);
158 qp = rp + dsize;
192 if (dp == rp || (quot && (dp == qp))) {
207 q_limb = mpihelp_divrem(qp,
[all...]
/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.c357 static int ehea_qp_alloc_register(struct ehea_qp *qp, struct hw_queue *hw_queue, argument
378 qp->fw_handle, rpage, 1);
402 struct ehea_qp *qp; local
407 qp = kzalloc(sizeof(*qp), GFP_KERNEL);
408 if (!qp)
411 qp->adapter = adapter;
414 &qp->fw_handle, &qp->epas);
425 ret = ehea_qp_alloc_register(qp,
490 ehea_destroy_qp_res(struct ehea_qp *qp, u64 force) argument
513 ehea_destroy_qp(struct ehea_qp *qp) argument
[all...]
/linux-master/drivers/infiniband/hw/qedr/
H A Dqedr_iw_cm.c84 struct qedr_qp *qp = container_of(ref, struct qedr_qp, refcnt); local
86 complete(&qp->qp_rel_comp);
94 if (ep->qp)
95 kref_put(&ep->qp->refcnt, qedr_iw_free_qp);
188 struct ib_qp *ibqp = &ep->qp->ibqp;
196 event.element.qp = ibqp;
215 struct qedr_qp *qp = ep->qp; local
218 /* The qp won't be released until we release the ep.
220 * function, therefore it is safe to access qp
521 struct qedr_qp *qp; local
543 struct qedr_qp *qp; local
741 struct qedr_qp *qp; local
802 struct qedr_qp *qp = get_qedr_qp(ibqp); local
809 struct qedr_qp *qp = get_qedr_qp(ibqp); local
[all...]
/linux-master/drivers/net/ethernet/qlogic/qed/
H A Dqed_rdma.c157 p_rdma_info->num_qps = num_cons / 2; /* 2 cids per qp */
1191 struct qed_rdma_qp *qp,
1197 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "icid = %08x\n", qp->icid);
1199 /* The following fields are filled in from qp and not FW as they can't
1202 out_params->mtu = qp->mtu;
1203 out_params->dest_qp = qp->dest_qp;
1204 out_params->incoming_atomic_en = qp->incoming_atomic_en;
1205 out_params->e2e_flow_control_en = qp->e2e_flow_control_en;
1206 out_params->incoming_rdma_read_en = qp->incoming_rdma_read_en;
1207 out_params->incoming_rdma_write_en = qp
1190 qed_rdma_query_qp(void *rdma_cxt, struct qed_rdma_qp *qp, struct qed_rdma_query_qp_out_params *out_params) argument
1230 qed_rdma_destroy_qp(void *rdma_cxt, struct qed_rdma_qp *qp) argument
1255 struct qed_rdma_qp *qp; local
1345 qed_rdma_modify_qp(void *rdma_cxt, struct qed_rdma_qp *qp, struct qed_rdma_modify_qp_in_params *params) argument
[all...]
H A Dqed_iwarp.h114 struct qed_rdma_qp *qp; member in struct:qed_iwarp_ep
176 struct qed_rdma_qp *qp,
179 int qed_iwarp_modify_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp,
182 int qed_iwarp_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp);
184 int qed_iwarp_fw_destroy(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp);
186 void qed_iwarp_query_qp(struct qed_rdma_qp *qp,
/linux-master/drivers/crypto/hisilicon/
H A Dqm.c845 static void qm_cq_head_update(struct hisi_qp *qp) argument
847 if (qp->qp_status.cq_head == qp->cq_depth - 1) {
848 qp->qp_status.cqc_phase = !qp->qp_status.cqc_phase;
849 qp->qp_status.cq_head = 0;
851 qp->qp_status.cq_head++;
855 static void qm_poll_req_cb(struct hisi_qp *qp) argument
857 struct qm_cqe *cqe = qp->cqe + qp
883 struct hisi_qp *qp; local
972 qm_set_qp_disable(struct hisi_qp *qp, int offset) argument
988 struct hisi_qp *qp = &qm->qp_array[qp_id]; local
1071 qm_init_qp_status(struct hisi_qp *qp) argument
1692 qm_stop_qp(struct hisi_qp *qp) argument
1808 qm_get_avail_sqe(struct hisi_qp *qp) argument
1819 hisi_qm_unset_hw_reset(struct hisi_qp *qp) argument
1831 struct hisi_qp *qp; local
1877 struct hisi_qp *qp; local
1900 hisi_qm_release_qp(struct hisi_qp *qp) argument
1914 qm_sq_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument
1940 qm_cq_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument
1969 qm_qp_ctx_cfg(struct hisi_qp *qp, int qp_id, u32 pasid) argument
1982 qm_start_qp_nolock(struct hisi_qp *qp, unsigned long arg) argument
2013 hisi_qm_start_qp(struct hisi_qp *qp, unsigned long arg) argument
2032 qp_stop_fail_cb(struct hisi_qp *qp) argument
2095 qm_drain_qp(struct hisi_qp *qp) argument
2130 qm_stop_qp_nolock(struct hisi_qp *qp) argument
2169 hisi_qm_stop_qp(struct hisi_qp *qp) argument
2192 hisi_qp_send(struct hisi_qp *qp, const void *msg) argument
2233 qm_qp_event_notifier(struct hisi_qp *qp) argument
2264 struct hisi_qp *qp; local
2283 struct hisi_qp *qp = q->priv; local
2293 struct hisi_qp *qp = q->priv; local
2344 struct hisi_qp *qp = q->priv; local
2351 struct hisi_qp *qp = q->priv; local
2380 struct hisi_qp *qp = q->priv; local
2398 struct hisi_qp *qp = q->priv; local
2408 struct hisi_qp *qp = q->priv; local
2746 struct hisi_qp *qp; local
3089 struct hisi_qp *qp; local
3119 struct hisi_qp *qp; local
3140 struct hisi_qp *qp; local
[all...]
/linux-master/drivers/infiniband/core/
H A Drestrack.c195 struct ib_qp *qp = container_of(res, struct ib_qp, res); local
197 WARN_ONCE(qp->qp_num >> 24 || qp->port >> 8,
198 "QP number 0x%0X and port 0x%0X", qp->qp_num,
199 qp->port);
200 res->id = qp->qp_num;
201 if (qp->qp_type == IB_QPT_SMI || qp->qp_type == IB_QPT_GSI)
202 res->id |= qp->port << 24;
207 if (qp
[all...]
H A Duverbs_std_types_qp.c15 struct ib_qp *qp = uobject->object; local
30 } else if (qp == qp->real_qp) {
31 ib_uverbs_detach_umcast(qp, uqp);
34 ret = ib_destroy_qp_user(qp, &attrs->driver_udata);
91 struct ib_qp *qp; local
251 qp = ib_create_qp_user(device, pd, &attr, &attrs->driver_udata, obj,
253 if (IS_ERR(qp)) {
254 ret = PTR_ERR(qp);
257 ib_qp_usecnt_inc(qp);
[all...]
/linux-master/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.c665 struct ocrdma_qp *qp)
670 if (qp == NULL)
672 ocrdma_qp_state_change(qp, new_ib_qps, &old_ib_qps);
678 struct ocrdma_qp *qp = NULL; local
691 * Some FW version returns wrong qp or cq ids in CQEs.
697 qp = dev->qp_tbl[qpid];
698 if (qp == NULL) {
733 ib_evt.element.qp = &qp->ibqp;
735 ocrdma_process_qpcat_error(dev, qp);
664 ocrdma_process_qpcat_error(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument
905 struct ocrdma_qp *qp; local
2078 ocrdma_is_qp_in_sq_flushlist(struct ocrdma_cq *cq, struct ocrdma_qp *qp) argument
2091 ocrdma_is_qp_in_rq_flushlist(struct ocrdma_cq *cq, struct ocrdma_qp *qp) argument
2104 ocrdma_flush_qp(struct ocrdma_qp *qp) argument
2122 ocrdma_init_hwq_ptr(struct ocrdma_qp *qp) argument
2130 ocrdma_qp_state_change(struct ocrdma_qp *qp, enum ib_qp_state new_ib_state, enum ib_qp_state *old_ib_state) argument
2161 ocrdma_set_create_qp_mbx_access_flags(struct ocrdma_qp *qp) argument
2177 ocrdma_set_create_qp_sq_cmd(struct ocrdma_create_qp_req *cmd, struct ib_qp_init_attr *attrs, struct ocrdma_qp *qp) argument
2232 ocrdma_set_create_qp_rq_cmd(struct ocrdma_create_qp_req *cmd, struct ib_qp_init_attr *attrs, struct ocrdma_qp *qp) argument
2279 ocrdma_set_create_qp_dpp_cmd(struct ocrdma_create_qp_req *cmd, struct ocrdma_pd *pd, struct ocrdma_qp *qp, u8 enable_dpp_cq, u16 dpp_cq_id) argument
2295 ocrdma_set_create_qp_ird_cmd(struct ocrdma_create_qp_req *cmd, struct ocrdma_qp *qp) argument
2328 ocrdma_get_create_qp_rsp(struct ocrdma_create_qp_rsp *rsp, struct ocrdma_qp *qp, struct ib_qp_init_attr *attrs, u16 *dpp_offset, u16 *dpp_credit_lmt) argument
2363 ocrdma_mbx_create_qp(struct ocrdma_qp *qp, struct ib_qp_init_attr *attrs, u8 enable_dpp_cq, u16 dpp_cq_id, u16 *dpp_offset, u16 *dpp_credit_lmt) argument
2462 ocrdma_mbx_query_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ocrdma_qp_params *param) argument
2483 ocrdma_set_av_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) argument
2569 ocrdma_set_qp_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) argument
2678 ocrdma_mbx_modify_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ib_qp_attr *attrs, int attr_mask) argument
2714 ocrdma_mbx_destroy_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument
[all...]
/linux-master/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c61 ret = ib_modify_qp(priv->qp, qp_attr, IB_QP_QKEY);
69 ret = ib_attach_mcast(priv->qp, mgid, mlid);
84 ret = ib_detach_mcast(priv->qp, mgid, mlid);
108 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask);
117 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask);
127 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask);
137 if (ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE))
212 priv->qp = ib_create_qp(priv->pd, &init_attr);
213 if (IS_ERR(priv->qp)) {
259 if (priv->qp) {
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Den_resources.c36 #include <linux/mlx4/qp.h>
78 /* disable multicast loopback to qp with same counter */
89 en_dbg(HW, priv, "Setting RX qp %x tunnel mode to RX tunneled & non-tunneled\n", qpn);
94 int mlx4_en_change_mcast_lb(struct mlx4_en_priv *priv, struct mlx4_qp *qp, argument
104 ret = mlx4_update_qp(priv->mdev->dev, qp->qpn,
111 void mlx4_en_sqp_event(struct mlx4_qp *qp, enum mlx4_event event) argument
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe.h52 #define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device, \
53 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
73 #define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \
74 "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
94 #define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp)
[all...]

Completed in 367 milliseconds

1234567891011>>