Searched refs:qp (Results 51 - 75 of 380) sorted by relevance

1234567891011>>

/linux-master/drivers/infiniband/hw/qedr/
H A Dqedr_iw_cm.h45 void qedr_iw_qp_add_ref(struct ib_qp *qp);
47 void qedr_iw_qp_rem_ref(struct ib_qp *qp);
H A Dqedr_roce_cm.h54 struct qedr_qp *qp);
56 struct qedr_qp *qp, struct ib_qp_init_attr *attrs);
/linux-master/drivers/infiniband/hw/mana/
H A DMakefile4 mana_ib-y := device.o main.o wq.o qp.o cq.o mr.o
/linux-master/drivers/infiniband/core/
H A Dcounters.c79 const struct ib_qp *qp,
88 param->qp_type = qp->qp_type;
92 struct ib_qp *qp)
96 if (qp->counter)
99 if (!qp->device->ops.counter_bind_qp)
103 ret = qp->device->ops.counter_bind_qp(counter, qp);
142 struct ib_qp *qp,
176 auto_mode_init_counter(counter, qp, port_counter->mode.mask);
191 ret = __rdma_counter_bind_qp(counter, qp);
78 auto_mode_init_counter(struct rdma_counter *counter, const struct ib_qp *qp, enum rdma_nl_counter_mask new_mask) argument
91 __rdma_counter_bind_qp(struct rdma_counter *counter, struct ib_qp *qp) argument
141 alloc_and_bind(struct ib_device *dev, u32 port, struct ib_qp *qp, enum rdma_nl_counter_mode mode) argument
225 auto_mode_match(struct ib_qp *qp, struct rdma_counter *counter, enum rdma_nl_counter_mask auto_mask) argument
241 __rdma_counter_unbind_qp(struct ib_qp *qp) argument
278 rdma_get_counter_auto_mode(struct ib_qp *qp, u32 port) argument
323 rdma_counter_bind_qp_auto(struct ib_qp *qp, u32 port) argument
361 rdma_counter_unbind_qp(struct ib_qp *qp, bool force) argument
447 struct ib_qp *qp = NULL; local
489 struct ib_qp *qp; local
539 struct ib_qp *qp; local
585 struct ib_qp *qp; local
[all...]
H A Dverbs.c1106 struct ib_qp *qp = context; local
1109 spin_lock_irqsave(&qp->device->qp_open_list_lock, flags);
1110 list_for_each_entry(event->element.qp, &qp->open_list, open_list)
1111 if (event->element.qp->event_handler)
1112 event->element.qp->event_handler(event, event->element.qp->qp_context);
1113 spin_unlock_irqrestore(&qp->device->qp_open_list_lock, flags);
1120 struct ib_qp *qp; local
1124 qp
1153 struct ib_qp *qp, *real_qp; local
1171 create_xrc_qp_user(struct ib_qp *qp, struct ib_qp_init_attr *qp_init_attr) argument
1206 struct ib_qp *qp; local
1282 struct ib_qp *qp, *xrc_qp; local
1302 ib_qp_usecnt_inc(struct ib_qp *qp) argument
1317 ib_qp_usecnt_dec(struct ib_qp *qp) argument
1337 struct ib_qp *qp; local
1744 is_qp_type_connected(const struct ib_qp *qp) argument
1755 _ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument
2005 ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask) argument
2013 ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument
2027 ib_close_qp(struct ib_qp *qp) argument
2049 __ib_destroy_shared_qp(struct ib_qp *qp) argument
2074 ib_destroy_qp_user(struct ib_qp *qp, struct ib_udata *udata) argument
2383 is_valid_mcast_lid(struct ib_qp *qp, u16 lid) argument
2421 ib_attach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument
2439 ib_detach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument
2817 __ib_drain_sq(struct ib_qp *qp) argument
2856 __ib_drain_rq(struct ib_qp *qp) argument
2906 ib_drain_sq(struct ib_qp *qp) argument
2935 ib_drain_rq(struct ib_qp *qp) argument
2960 ib_drain_qp(struct ib_qp *qp) argument
[all...]
H A Dsecurity.c164 * the qp pointer in the security structure is
170 ib_modify_qp(sec->qp,
174 if (sec->qp->event_handler && sec->qp->qp_context) {
175 event.element.qp = sec->qp;
176 sec->qp->event_handler(&event,
177 sec->qp->qp_context);
183 struct ib_qp *qp = shared_qp_sec->qp; local
331 get_new_pps(const struct ib_qp *qp, const struct ib_qp_attr *qp_attr, int qp_attr_mask) argument
373 ib_open_shared_qp_security(struct ib_qp *qp, struct ib_device *dev) argument
415 ib_create_qp_security(struct ib_qp *qp, struct ib_device *dev) argument
571 ib_security_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_udata *udata) argument
[all...]
/linux-master/include/rdma/
H A Drw.h45 int rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num,
48 void rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
52 int rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
57 void rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
62 struct ib_send_wr *rdma_rw_ctx_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
64 int rdma_rw_ctx_post(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num,
70 int rdma_rw_init_mrs(struct ib_qp *qp, struct ib_qp_init_attr *attr);
71 void rdma_rw_cleanup_mrs(struct ib_qp *qp);
/linux-master/drivers/infiniband/hw/mlx5/
H A Dwr.c7 #include <linux/mlx5/qp.h>
54 static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, argument
88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size,
260 static int set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, argument
278 if (unlikely(inl > qp->max_inline_data))
285 handle_post_send_edge(&qp->sq, wqe,
419 struct mlx5_ib_qp *qp, void **seg, int *size,
514 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
523 handle_post_send_edge(&qp
416 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument
557 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument
635 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) argument
692 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument
705 dump_wqe(struct mlx5_ib_qp *qp, u32 idx, int size_16) argument
724 mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, unsigned int *idx, int *size, void **cur_edge, int nreq, __be32 general_id, bool send_signaled, bool solicited) argument
748 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) argument
758 mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) argument
794 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) argument
803 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) argument
812 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) argument
843 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) argument
931 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) argument
994 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) argument
1004 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) argument
1025 mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq, struct mlx5_wqe_ctrl_seg *ctrl) argument
1057 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
1209 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
[all...]
H A Dqp.c44 #include "qp.h"
80 struct mlx5_core_qp *qp; member in struct:mlx5_ib_qp_event_work
146 static int mlx5_ib_read_kernel_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, argument
155 wqe_index = wqe_index & qp->sq.fbc.sz_m1;
158 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index);
174 wqe_index = (wqe_index + 1) & qp->sq.fbc.sz_m1;
175 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index);
181 static int mlx5_ib_read_user_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, argument
184 struct mlx5_ib_qp_base *base = &qp->trans_qp.base;
186 struct mlx5_ib_wq *wq = &qp
232 mlx5_ib_read_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument
248 mlx5_ib_read_user_wqe_rq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument
268 mlx5_ib_read_wqe_rq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument
321 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
406 mlx5_ib_qp_event(struct mlx5_core_qp *qp, int type) argument
433 set_rq_size(struct mlx5_ib_dev *dev, struct ib_qp_cap *cap, int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) argument
591 calc_sq_size(struct mlx5_ib_dev *dev, struct ib_qp_init_attr *attr, struct mlx5_ib_qp *qp) argument
636 set_user_buf_size(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd, struct mlx5_ib_qp_base *base, struct ib_qp_init_attr *attr) argument
941 _create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct ib_udata *udata, struct ib_qp_init_attr *attr, u32 **in, struct mlx5_ib_create_qp_resp *resp, int *inlen, struct mlx5_ib_qp_base *base, struct mlx5_ib_create_qp *ucmd) argument
1067 destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_qp_base *base, struct ib_udata *udata) argument
1099 _create_kernel_qp(struct mlx5_ib_dev *dev, struct ib_qp_init_attr *init_attr, struct mlx5_ib_qp *qp, u32 **in, int *inlen, struct mlx5_ib_qp_base *base) argument
1219 get_rx_type(struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument
1230 create_raw_packet_qp_tis(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_sq *sq, u32 tdn, struct ib_pd *pd) argument
1569 create_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, u32 *in, size_t inlen, struct ib_pd *pd, struct ib_udata *udata, struct mlx5_ib_create_qp_resp *resp, struct ib_qp_init_attr *init_attr) argument
1664 destroy_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument
1682 raw_packet_qp_copy_info(struct mlx5_ib_qp *qp, struct mlx5_ib_raw_packet_qp *raw_packet_qp) argument
1694 destroy_rss_raw_qp_tir(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument
1715 create_rss_raw_qp_tir(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
1921 configure_requester_scat_cqe(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *init_attr, void *qpc) argument
1992 create_xrc_tgt_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
2068 create_dci(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
2229 create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
2432 create_kernel_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
2658 destroy_qp_common(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_udata *udata) argument
2729 create_dct(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
2844 process_vendor_flag(struct mlx5_ib_dev *dev, int *flags, int flag, bool cond, struct mlx5_ib_qp *qp) argument
2872 process_vendor_flags(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, void *ucmd, struct ib_qp_init_attr *attr) argument
2950 process_create_flag(struct mlx5_ib_dev *dev, int *flags, int flag, bool cond, struct mlx5_ib_qp *qp) argument
2973 process_create_flags(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument
3086 create_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
3146 check_qp_attr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument
3173 get_qp_uidx(struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument
3244 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
3335 mlx5_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata) argument
3350 set_qpc_atomic_flags(struct mlx5_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask, void *qpc) argument
3515 mlx5_set_path(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct rdma_ah_attr *ah, void *path, u8 port, int attr_mask, u32 path_flags, const struct ib_qp_attr *attr, bool alt) argument
3897 modify_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct mlx5_modify_raw_qp_param *raw_qp_param, u8 tx_affinity) argument
4001 qp_supports_affinity(struct mlx5_ib_qp *qp) argument
4011 get_tx_affinity(struct ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask, u8 init, struct ib_udata *udata) argument
4047 __mlx5_ib_qp_set_raw_qp_counter(struct mlx5_ib_qp *qp, u32 set_id, struct mlx5_core_dev *mdev) argument
4071 __mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter) argument
4148 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
4494 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
4603 mlx5_ib_modify_qp_allowed(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument
4658 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
4860 sqrq_state_to_qp_state(u8 sq_state, u8 rq_state, struct mlx5_ib_qp *qp, u8 *qp_state) argument
4905 query_raw_packet_qp_state(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, u8 *raw_packet_qp_state) argument
4932 query_qp_attr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_attr *qp_attr) argument
5065 struct mlx5_ib_qp *qp = to_mqp(ibqp); local
5736 mlx5_ib_drain_sq(struct ib_qp *qp) argument
5771 mlx5_ib_drain_rq(struct ib_qp *qp) argument
5805 mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter) argument
[all...]
H A Dqp.h24 int mlx5_core_create_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *qp,
26 int mlx5_qpc_create_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp,
29 void *qpc, struct mlx5_core_qp *qp, u32 *ece);
30 int mlx5_core_destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp);
32 int mlx5_core_qp_query(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp,
56 int mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter);
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_qp_rx.c330 static struct siw_wqe *siw_rqe_get(struct siw_qp *qp) argument
338 srq = qp->srq;
346 if (unlikely(!qp->recvq))
349 rqe = &qp->recvq[qp->rq_get % qp->attrs.rq_size];
357 wqe = rx_wqe(&qp->rx_untagged);
377 siw_dbg_qp(qp, "too many sge's: %d\n", rqe->num_sge);
383 qp->rq_get++;
435 int siw_proc_send(struct siw_qp *qp) argument
545 siw_proc_write(struct siw_qp *qp) argument
627 siw_proc_rreq(struct siw_qp *qp) argument
656 siw_init_rresp(struct siw_qp *qp, struct siw_rx_stream *srx) argument
739 siw_orqe_start_rx(struct siw_qp *qp) argument
782 siw_proc_rresp(struct siw_qp *qp) argument
878 siw_proc_terminate(struct siw_qp *qp) argument
951 siw_get_trailer(struct siw_qp *qp, struct siw_rx_stream *srx) argument
998 struct siw_qp *qp = rx_qp(srx); local
1130 siw_check_tx_fence(struct siw_qp *qp) argument
1199 siw_rdmap_complete(struct siw_qp *qp, int error) argument
1339 struct siw_qp *qp = rd_desc->arg.data; local
[all...]
H A Dsiw_verbs.c296 * @qp: Queue pait
305 struct siw_qp *qp = to_siw_qp(ibqp); local
359 init_rwsem(&qp->state_lock);
360 spin_lock_init(&qp->sq_lock);
361 spin_lock_init(&qp->rq_lock);
362 spin_lock_init(&qp->orq_lock);
364 rv = siw_qp_add(sdev, qp);
380 qp->sendq = vmalloc_user(num_sqe * sizeof(struct siw_sqe));
382 qp->sendq = vcalloc(num_sqe, sizeof(struct siw_sqe));
384 if (qp
507 struct siw_qp *qp; local
545 struct siw_qp *qp = to_siw_qp(base_qp); local
591 struct siw_qp *qp = to_siw_qp(base_qp); local
675 siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument
728 siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument
759 struct siw_qp *qp = to_siw_qp(base_qp); local
999 struct siw_qp *qp = to_siw_qp(base_qp); local
1809 siw_qp_event(struct siw_qp *qp, enum ib_event_type etype) argument
[all...]
/linux-master/drivers/interconnect/qcom/
H A Dicc-rpmh.c25 struct qcom_icc_provider *qp; local
28 qp = to_qcom_provider(node->provider);
36 qcom_icc_bcm_voter_add(qp->voter, qn->bcms[i]);
88 struct qcom_icc_provider *qp; local
96 qp = to_qcom_provider(node->provider);
98 qcom_icc_bcm_voter_commit(qp->voter);
169 struct qcom_icc_provider *qp; local
181 qp = devm_kzalloc(dev, sizeof(*qp), GFP_KERNEL);
182 if (!qp)
258 struct qcom_icc_provider *qp = platform_get_drvdata(pdev); local
[all...]
/linux-master/drivers/scsi/bnx2i/
H A Dbnx2i_hwi.c153 cq_db = (struct bnx2i_5771x_cq_db *) ep->qp.cq_pgtbl_virt;
170 cq_index = ep->qp.cqe_exp_seq_sn + next_index - 1;
171 if (cq_index > ep->qp.cqe_size * 2)
172 cq_index -= ep->qp.cqe_size * 2;
195 if (!bnx2i_conn->ep->qp.rqe_left)
198 bnx2i_conn->ep->qp.rqe_left--;
199 memcpy(ptr, (u8 *) bnx2i_conn->ep->qp.rq_cons_qe, len);
200 if (bnx2i_conn->ep->qp.rq_cons_qe == bnx2i_conn->ep->qp.rq_last_qe) {
201 bnx2i_conn->ep->qp
1960 struct qp_info *qp; local
[all...]
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_mw.c50 static int rxe_check_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, argument
76 if (unlikely(qp->pd != to_rpd(mw->ibmw.pd))) {
78 "attempt to bind type 2 MW with qp with different PD\n");
135 static void rxe_do_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, argument
159 rxe_get(qp);
160 mw->qp = qp;
164 int rxe_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe) argument
169 struct rxe_dev *rxe = to_rdev(qp->ibqp.device);
208 ret = rxe_check_bind_mw(qp, wq
224 rxe_check_invalidate_mw(struct rxe_qp *qp, struct rxe_mw *mw) argument
238 struct rxe_qp *qp; local
258 rxe_invalidate_mw(struct rxe_qp *qp, u32 rkey) argument
290 rxe_lookup_mw(struct rxe_qp *qp, int access, u32 rkey) argument
328 struct rxe_qp *qp = mw->qp; local
[all...]
H A Drxe_av.c22 struct rxe_qp *qp; local
30 qp = obj;
31 rxe = to_rdev(qp->ibqp.device);
42 rxe_dbg_qp(qp, "invalid sgid index = %d\n",
54 rxe_dbg_qp(qp, "invalid network type for rdma_rxe = %d\n",
63 int rxe_av_chk_attr(struct rxe_qp *qp, struct rdma_ah_attr *attr) argument
65 return chk_attr(qp, attr, false);
139 if (!pkt || !pkt->qp)
142 if (qp_type(pkt->qp) == IB_QPT_RC || qp_type(pkt->qp)
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dipoib_main.c57 struct rvt_qp *qp; local
61 qp = rvt_lookup_qpn(ib_to_rvt(priv->device), &ibp->rvp, qpn);
62 if (!qp) {
67 rvt_get_qp(qp);
68 priv->qp = qp;
82 if (!priv->qp)
88 rvt_put_qp(priv->qp);
89 priv->qp = NULL;
111 struct rvt_qp *qp; local
141 struct rvt_qp *qp; local
[all...]
/linux-master/drivers/interconnect/
H A Dicc-clk.c81 struct icc_clk_provider *qp; local
91 qp = devm_kzalloc(dev, struct_size(qp, clocks, num_clocks), GFP_KERNEL);
92 if (!qp)
95 qp->num_clocks = num_clocks;
97 provider = &qp->provider;
109 qp->clocks[i].clk = data[i].clk;
118 node->data = &qp->clocks[i];
157 struct icc_clk_provider *qp = container_of(provider, struct icc_clk_provider, provider); local
160 icc_provider_deregister(&qp
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c110 if (unlikely(conn->qp.rq.pc - conn->qp.rq.cc >= conn->qp.rq.size)) {
115 ix = conn->qp.rq.pc & (conn->qp.rq.size - 1);
116 data = mlx5_wq_cyc_get_wqe(&conn->qp.wq.rq, ix);
121 conn->qp.rq.pc++;
122 conn->qp.rq.bufs[ix] = buf;
126 *conn->qp.wq.rq.db = cpu_to_be32(conn->qp
[all...]
/linux-master/drivers/infiniband/hw/qib/
H A Dqib_verbs.c216 * @qp: the QP the packet came on
223 int has_grh, void *data, u32 tlen, struct rvt_qp *qp)
227 spin_lock(&qp->r_lock);
230 if (!(ib_rvt_state_ops[qp->state] & RVT_PROCESS_RECV_OK)) {
235 switch (qp->ibqp.qp_type) {
242 qib_ud_rcv(ibp, hdr, has_grh, data, tlen, qp);
246 qib_rc_rcv(rcd, hdr, has_grh, data, tlen, qp);
250 qib_uc_rcv(ibp, hdr, has_grh, data, tlen, qp);
258 spin_unlock(&qp->r_lock);
279 struct rvt_qp *qp; local
222 qib_qp_rcv(struct qib_ctxtdata *rcd, struct ib_header *hdr, int has_grh, void *data, u32 tlen, struct rvt_qp *qp) argument
366 struct rvt_qp *qp = NULL; local
559 __get_txreq(struct qib_ibdev *dev, struct rvt_qp *qp) argument
591 get_txreq(struct qib_ibdev *dev, struct rvt_qp *qp) argument
616 struct rvt_qp *qp; local
669 struct rvt_qp *qp; local
715 struct rvt_qp *qp = tx->qp; local
746 wait_kmem(struct qib_ibdev *dev, struct rvt_qp *qp) argument
770 qib_verbs_send_dma(struct rvt_qp *qp, struct ib_header *hdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
874 no_bufs_available(struct rvt_qp *qp) argument
906 qib_verbs_send_pio(struct rvt_qp *qp, struct ib_header *ibhdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
1014 qib_verbs_send(struct rvt_qp *qp, struct ib_header *hdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len) argument
1148 struct rvt_qp *qp; local
1683 _qib_schedule_send(struct rvt_qp *qp) argument
1700 qib_schedule_send(struct rvt_qp *qp) argument
[all...]
H A Dqib_verbs.h218 static inline int qib_send_ok(struct rvt_qp *qp) argument
220 return !(qp->s_flags & (RVT_S_BUSY | RVT_S_ANY_WAIT_IO)) &&
221 (qp->s_hdrwords || (qp->s_flags & RVT_S_RESP_PENDING) ||
222 !(qp->s_flags & RVT_S_ANY_WAIT_SEND));
225 bool _qib_schedule_send(struct rvt_qp *qp);
226 bool qib_schedule_send(struct rvt_qp *qp);
272 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp);
273 void qib_qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp);
274 void qib_notify_qp_reset(struct rvt_qp *qp);
[all...]
/linux-master/drivers/scsi/qla2xxx/
H A Dqla_mid.c1084 int qla_create_buf_pool(struct scsi_qla_host *vha, struct qla_qpair *qp) argument
1088 qp->buf_pool.num_bufs = qp->req->length;
1090 sz = BITS_TO_LONGS(qp->req->length);
1091 qp->buf_pool.buf_map = kcalloc(sz, sizeof(long), GFP_KERNEL);
1092 if (!qp->buf_pool.buf_map) {
1097 sz = qp->req->length * sizeof(void *);
1098 qp->buf_pool.buf_array = kcalloc(qp->req->length, sizeof(void *), GFP_KERNEL);
1099 if (!qp
1118 qla_free_buf_pool(struct qla_qpair *qp) argument
1137 qla_get_buf(struct scsi_qla_host *vha, struct qla_qpair *qp, struct qla_buf_dsc *dsc) argument
1189 qla_trim_buf(struct qla_qpair *qp, u16 trim) argument
1221 __qla_adjust_buf(struct qla_qpair *qp) argument
1243 qla_put_buf(struct qla_qpair *qp, struct qla_buf_dsc *dsc) argument
1262 struct qla_qpair *qp; local
[all...]
/linux-master/drivers/net/ethernet/qlogic/qed/
H A Dqed_roce.h23 int qed_roce_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp);
26 struct qed_rdma_qp *qp,
30 struct qed_rdma_qp *qp,
/linux-master/drivers/net/
H A Dntb_netdev.c76 struct ntb_transport_qp *qp; member in struct:ntb_netdev
89 ntb_transport_link_query(dev->qp));
92 if (ntb_transport_link_query(dev->qp))
99 static void ntb_netdev_rx_handler(struct ntb_transport_qp *qp, void *qp_data, argument
138 rc = ntb_transport_rx_enqueue(qp, skb, skb->data, ndev->mtu + ETH_HLEN);
147 struct ntb_transport_qp *qp, int size)
157 if (likely(ntb_transport_tx_free_entry(qp) < size)) {
167 struct ntb_transport_qp *qp, int size)
170 (ntb_transport_tx_free_entry(qp) >= size))
173 return __ntb_netdev_maybe_stop_tx(ndev, qp, siz
146 __ntb_netdev_maybe_stop_tx(struct net_device *netdev, struct ntb_transport_qp *qp, int size) argument
166 ntb_netdev_maybe_stop_tx(struct net_device *ndev, struct ntb_transport_qp *qp, int size) argument
176 ntb_netdev_tx_handler(struct ntb_transport_qp *qp, void *qp_data, void *data, int len) argument
[all...]
/linux-master/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c1081 static int ocrdma_add_qpn_map(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument
1085 if (qp->id < OCRDMA_MAX_QP && dev->qp_tbl[qp->id] == NULL) {
1086 dev->qp_tbl[qp->id] = qp;
1092 static void ocrdma_del_qpn_map(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument
1094 dev->qp_tbl[qp->id] = NULL;
1105 pr_err("%s(%d) unsupported qp type=0x%x requested\n",
1171 static int ocrdma_copy_qp_uresp(struct ocrdma_qp *qp, argument
1178 struct ocrdma_pd *pd = qp
1230 ocrdma_set_qp_db(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ocrdma_pd *pd) argument
1250 ocrdma_alloc_wr_id_tbl(struct ocrdma_qp *qp) argument
1265 ocrdma_set_qp_init_params(struct ocrdma_qp *qp, struct ocrdma_pd *pd, struct ib_qp_init_attr *attrs) argument
1299 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local
1368 struct ocrdma_qp *qp; local
1389 struct ocrdma_qp *qp; local
1462 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local
1556 is_hw_sq_empty(struct ocrdma_qp *qp) argument
1561 is_hw_rq_empty(struct ocrdma_qp *qp) argument
1588 ocrdma_discard_cqes(struct ocrdma_qp *qp, struct ocrdma_cq *cq) argument
1650 ocrdma_del_flush_qp(struct ocrdma_qp *qp) argument
1672 struct ocrdma_qp *qp; local
1872 ocrdma_build_ud_hdr(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument
1918 ocrdma_build_inline_sges(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, struct ocrdma_sge *sge, const struct ib_send_wr *wr, u32 wqe_size) argument
1958 ocrdma_build_send(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument
1975 ocrdma_build_write(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument
1993 ocrdma_build_read(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument
2022 ocrdma_build_reg(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_reg_wr *wr) argument
2078 ocrdma_ring_sq_db(struct ocrdma_qp *qp) argument
2089 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local
2190 ocrdma_ring_rq_db(struct ocrdma_qp *qp) argument
2223 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local
2399 ocrdma_update_wc(struct ocrdma_qp *qp, struct ib_wc *ibwc, u32 wqe_idx) argument
2437 ocrdma_set_cqe_status_flushed(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe) argument
2469 ocrdma_update_err_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument
2491 ocrdma_update_err_rcqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument
2501 ocrdma_update_err_scqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument
2511 ocrdma_poll_err_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument
2555 ocrdma_poll_success_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled) argument
2581 ocrdma_poll_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument
2624 ocrdma_update_free_srq_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp) argument
2644 ocrdma_poll_err_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop, int status) argument
2679 ocrdma_poll_success_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc) argument
2713 ocrdma_poll_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument
2758 struct ocrdma_qp *qp = NULL; local
2811 ocrdma_add_err_cqe(struct ocrdma_cq *cq, int num_entries, struct ocrdma_qp *qp, struct ib_wc *ibwc) argument
2843 struct ocrdma_qp *qp; local
[all...]

Completed in 263 milliseconds

1234567891011>>