/linux-master/drivers/infiniband/hw/qedr/ |
H A D | qedr_iw_cm.h | 45 void qedr_iw_qp_add_ref(struct ib_qp *qp); 47 void qedr_iw_qp_rem_ref(struct ib_qp *qp);
|
H A D | qedr_roce_cm.h | 54 struct qedr_qp *qp); 56 struct qedr_qp *qp, struct ib_qp_init_attr *attrs);
|
/linux-master/drivers/infiniband/hw/mana/ |
H A D | Makefile | 4 mana_ib-y := device.o main.o wq.o qp.o cq.o mr.o
|
/linux-master/drivers/infiniband/core/ |
H A D | counters.c | 79 const struct ib_qp *qp, 88 param->qp_type = qp->qp_type; 92 struct ib_qp *qp) 96 if (qp->counter) 99 if (!qp->device->ops.counter_bind_qp) 103 ret = qp->device->ops.counter_bind_qp(counter, qp); 142 struct ib_qp *qp, 176 auto_mode_init_counter(counter, qp, port_counter->mode.mask); 191 ret = __rdma_counter_bind_qp(counter, qp); 78 auto_mode_init_counter(struct rdma_counter *counter, const struct ib_qp *qp, enum rdma_nl_counter_mask new_mask) argument 91 __rdma_counter_bind_qp(struct rdma_counter *counter, struct ib_qp *qp) argument 141 alloc_and_bind(struct ib_device *dev, u32 port, struct ib_qp *qp, enum rdma_nl_counter_mode mode) argument 225 auto_mode_match(struct ib_qp *qp, struct rdma_counter *counter, enum rdma_nl_counter_mask auto_mask) argument 241 __rdma_counter_unbind_qp(struct ib_qp *qp) argument 278 rdma_get_counter_auto_mode(struct ib_qp *qp, u32 port) argument 323 rdma_counter_bind_qp_auto(struct ib_qp *qp, u32 port) argument 361 rdma_counter_unbind_qp(struct ib_qp *qp, bool force) argument 447 struct ib_qp *qp = NULL; local 489 struct ib_qp *qp; local 539 struct ib_qp *qp; local 585 struct ib_qp *qp; local [all...] |
H A D | verbs.c | 1106 struct ib_qp *qp = context; local 1109 spin_lock_irqsave(&qp->device->qp_open_list_lock, flags); 1110 list_for_each_entry(event->element.qp, &qp->open_list, open_list) 1111 if (event->element.qp->event_handler) 1112 event->element.qp->event_handler(event, event->element.qp->qp_context); 1113 spin_unlock_irqrestore(&qp->device->qp_open_list_lock, flags); 1120 struct ib_qp *qp; local 1124 qp 1153 struct ib_qp *qp, *real_qp; local 1171 create_xrc_qp_user(struct ib_qp *qp, struct ib_qp_init_attr *qp_init_attr) argument 1206 struct ib_qp *qp; local 1282 struct ib_qp *qp, *xrc_qp; local 1302 ib_qp_usecnt_inc(struct ib_qp *qp) argument 1317 ib_qp_usecnt_dec(struct ib_qp *qp) argument 1337 struct ib_qp *qp; local 1744 is_qp_type_connected(const struct ib_qp *qp) argument 1755 _ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 2005 ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask) argument 2013 ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument 2027 ib_close_qp(struct ib_qp *qp) argument 2049 __ib_destroy_shared_qp(struct ib_qp *qp) argument 2074 ib_destroy_qp_user(struct ib_qp *qp, struct ib_udata *udata) argument 2383 is_valid_mcast_lid(struct ib_qp *qp, u16 lid) argument 2421 ib_attach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument 2439 ib_detach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument 2817 __ib_drain_sq(struct ib_qp *qp) argument 2856 __ib_drain_rq(struct ib_qp *qp) argument 2906 ib_drain_sq(struct ib_qp *qp) argument 2935 ib_drain_rq(struct ib_qp *qp) argument 2960 ib_drain_qp(struct ib_qp *qp) argument [all...] |
H A D | security.c | 164 * the qp pointer in the security structure is 170 ib_modify_qp(sec->qp, 174 if (sec->qp->event_handler && sec->qp->qp_context) { 175 event.element.qp = sec->qp; 176 sec->qp->event_handler(&event, 177 sec->qp->qp_context); 183 struct ib_qp *qp = shared_qp_sec->qp; local 331 get_new_pps(const struct ib_qp *qp, const struct ib_qp_attr *qp_attr, int qp_attr_mask) argument 373 ib_open_shared_qp_security(struct ib_qp *qp, struct ib_device *dev) argument 415 ib_create_qp_security(struct ib_qp *qp, struct ib_device *dev) argument 571 ib_security_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_udata *udata) argument [all...] |
/linux-master/include/rdma/ |
H A D | rw.h | 45 int rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, 48 void rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp, 52 int rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, 57 void rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp, 62 struct ib_send_wr *rdma_rw_ctx_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, 64 int rdma_rw_ctx_post(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, 70 int rdma_rw_init_mrs(struct ib_qp *qp, struct ib_qp_init_attr *attr); 71 void rdma_rw_cleanup_mrs(struct ib_qp *qp);
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | wr.c | 7 #include <linux/mlx5/qp.h> 54 static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, argument 88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); 91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, 260 static int set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, argument 278 if (unlikely(inl > qp->max_inline_data)) 285 handle_post_send_edge(&qp->sq, wqe, 419 struct mlx5_ib_qp *qp, void **seg, int *size, 514 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); 523 handle_post_send_edge(&qp 416 set_sig_data_segment(const struct ib_send_wr *send_wr, struct ib_mr *sig_mr, struct ib_sig_attrs *sig_attrs, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument 557 set_pi_umr_wr(const struct ib_send_wr *send_wr, struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument 635 set_reg_wr(struct mlx5_ib_qp *qp, const struct ib_reg_wr *wr, void **seg, int *size, void **cur_edge, bool check_not_free) argument 692 set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size, void **cur_edge) argument 705 dump_wqe(struct mlx5_ib_qp *qp, u32 idx, int size_16) argument 724 mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, unsigned int *idx, int *size, void **cur_edge, int nreq, __be32 general_id, bool send_signaled, bool solicited) argument 748 begin_wqe(struct mlx5_ib_qp *qp, void **seg, struct mlx5_wqe_ctrl_seg **ctrl, const struct ib_send_wr *wr, unsigned int *idx, int *size, void **cur_edge, int nreq) argument 758 mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, void *seg, u8 size, void *cur_edge, unsigned int idx, u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) argument 794 handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) argument 803 handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int idx) argument 812 handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, struct ib_sig_domain *domain, u32 psv_index, u8 next_fence) argument 843 handle_reg_mr_integrity(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence) argument 931 handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size, void **cur_edge, unsigned int *idx, int nreq, u8 fence, u8 next_fence, int *num_sge) argument 994 handle_qpt_hw_gsi(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) argument 1004 handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, void **seg, int *size, void **cur_edge) argument 1025 mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq, struct mlx5_wqe_ctrl_seg *ctrl) argument 1057 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 1209 struct mlx5_ib_qp *qp = to_mqp(ibqp); local [all...] |
H A D | qp.c | 44 #include "qp.h" 80 struct mlx5_core_qp *qp; member in struct:mlx5_ib_qp_event_work 146 static int mlx5_ib_read_kernel_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, argument 155 wqe_index = wqe_index & qp->sq.fbc.sz_m1; 158 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index); 174 wqe_index = (wqe_index + 1) & qp->sq.fbc.sz_m1; 175 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index); 181 static int mlx5_ib_read_user_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, argument 184 struct mlx5_ib_qp_base *base = &qp->trans_qp.base; 186 struct mlx5_ib_wq *wq = &qp 232 mlx5_ib_read_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument 248 mlx5_ib_read_user_wqe_rq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument 268 mlx5_ib_read_wqe_rq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, size_t buflen, size_t *bc) argument 321 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 406 mlx5_ib_qp_event(struct mlx5_core_qp *qp, int type) argument 433 set_rq_size(struct mlx5_ib_dev *dev, struct ib_qp_cap *cap, int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) argument 591 calc_sq_size(struct mlx5_ib_dev *dev, struct ib_qp_init_attr *attr, struct mlx5_ib_qp *qp) argument 636 set_user_buf_size(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd, struct mlx5_ib_qp_base *base, struct ib_qp_init_attr *attr) argument 941 _create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct ib_udata *udata, struct ib_qp_init_attr *attr, u32 **in, struct mlx5_ib_create_qp_resp *resp, int *inlen, struct mlx5_ib_qp_base *base, struct mlx5_ib_create_qp *ucmd) argument 1067 destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_qp_base *base, struct ib_udata *udata) argument 1099 _create_kernel_qp(struct mlx5_ib_dev *dev, struct ib_qp_init_attr *init_attr, struct mlx5_ib_qp *qp, u32 **in, int *inlen, struct mlx5_ib_qp_base *base) argument 1219 get_rx_type(struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument 1230 create_raw_packet_qp_tis(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_sq *sq, u32 tdn, struct ib_pd *pd) argument 1569 create_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, u32 *in, size_t inlen, struct ib_pd *pd, struct ib_udata *udata, struct mlx5_ib_create_qp_resp *resp, struct ib_qp_init_attr *init_attr) argument 1664 destroy_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument 1682 raw_packet_qp_copy_info(struct mlx5_ib_qp *qp, struct mlx5_ib_raw_packet_qp *raw_packet_qp) argument 1694 destroy_rss_raw_qp_tir(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument 1715 create_rss_raw_qp_tir(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 1921 configure_requester_scat_cqe(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *init_attr, void *qpc) argument 1992 create_xrc_tgt_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 2068 create_dci(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 2229 create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 2432 create_kernel_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 2658 destroy_qp_common(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_udata *udata) argument 2729 create_dct(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 2844 process_vendor_flag(struct mlx5_ib_dev *dev, int *flags, int flag, bool cond, struct mlx5_ib_qp *qp) argument 2872 process_vendor_flags(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, void *ucmd, struct ib_qp_init_attr *attr) argument 2950 process_create_flag(struct mlx5_ib_dev *dev, int *flags, int flag, bool cond, struct mlx5_ib_qp *qp) argument 2973 process_create_flags(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument 3086 create_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 3146 check_qp_attr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_init_attr *attr) argument 3173 get_qp_uidx(struct mlx5_ib_qp *qp, struct mlx5_create_qp_params *params) argument 3244 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 3335 mlx5_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata) argument 3350 set_qpc_atomic_flags(struct mlx5_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask, void *qpc) argument 3515 mlx5_set_path(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct rdma_ah_attr *ah, void *path, u8 port, int attr_mask, u32 path_flags, const struct ib_qp_attr *attr, bool alt) argument 3897 modify_raw_packet_qp(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, const struct mlx5_modify_raw_qp_param *raw_qp_param, u8 tx_affinity) argument 4001 qp_supports_affinity(struct mlx5_ib_qp *qp) argument 4011 get_tx_affinity(struct ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask, u8 init, struct ib_udata *udata) argument 4047 __mlx5_ib_qp_set_raw_qp_counter(struct mlx5_ib_qp *qp, u32 set_id, struct mlx5_core_dev *mdev) argument 4071 __mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter) argument 4148 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 4494 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 4603 mlx5_ib_modify_qp_allowed(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp) argument 4658 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 4860 sqrq_state_to_qp_state(u8 sq_state, u8 rq_state, struct mlx5_ib_qp *qp, u8 *qp_state) argument 4905 query_raw_packet_qp_state(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, u8 *raw_packet_qp_state) argument 4932 query_qp_attr(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct ib_qp_attr *qp_attr) argument 5065 struct mlx5_ib_qp *qp = to_mqp(ibqp); local 5736 mlx5_ib_drain_sq(struct ib_qp *qp) argument 5771 mlx5_ib_drain_rq(struct ib_qp *qp) argument 5805 mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter) argument [all...] |
H A D | qp.h | 24 int mlx5_core_create_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *qp, 26 int mlx5_qpc_create_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, 29 void *qpc, struct mlx5_core_qp *qp, u32 *ece); 30 int mlx5_core_destroy_qp(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp); 32 int mlx5_core_qp_query(struct mlx5_ib_dev *dev, struct mlx5_core_qp *qp, 56 int mlx5_ib_qp_set_counter(struct ib_qp *qp, struct rdma_counter *counter);
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_qp_rx.c | 330 static struct siw_wqe *siw_rqe_get(struct siw_qp *qp) argument 338 srq = qp->srq; 346 if (unlikely(!qp->recvq)) 349 rqe = &qp->recvq[qp->rq_get % qp->attrs.rq_size]; 357 wqe = rx_wqe(&qp->rx_untagged); 377 siw_dbg_qp(qp, "too many sge's: %d\n", rqe->num_sge); 383 qp->rq_get++; 435 int siw_proc_send(struct siw_qp *qp) argument 545 siw_proc_write(struct siw_qp *qp) argument 627 siw_proc_rreq(struct siw_qp *qp) argument 656 siw_init_rresp(struct siw_qp *qp, struct siw_rx_stream *srx) argument 739 siw_orqe_start_rx(struct siw_qp *qp) argument 782 siw_proc_rresp(struct siw_qp *qp) argument 878 siw_proc_terminate(struct siw_qp *qp) argument 951 siw_get_trailer(struct siw_qp *qp, struct siw_rx_stream *srx) argument 998 struct siw_qp *qp = rx_qp(srx); local 1130 siw_check_tx_fence(struct siw_qp *qp) argument 1199 siw_rdmap_complete(struct siw_qp *qp, int error) argument 1339 struct siw_qp *qp = rd_desc->arg.data; local [all...] |
H A D | siw_verbs.c | 296 * @qp: Queue pait 305 struct siw_qp *qp = to_siw_qp(ibqp); local 359 init_rwsem(&qp->state_lock); 360 spin_lock_init(&qp->sq_lock); 361 spin_lock_init(&qp->rq_lock); 362 spin_lock_init(&qp->orq_lock); 364 rv = siw_qp_add(sdev, qp); 380 qp->sendq = vmalloc_user(num_sqe * sizeof(struct siw_sqe)); 382 qp->sendq = vcalloc(num_sqe, sizeof(struct siw_sqe)); 384 if (qp 507 struct siw_qp *qp; local 545 struct siw_qp *qp = to_siw_qp(base_qp); local 591 struct siw_qp *qp = to_siw_qp(base_qp); local 675 siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 728 siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 759 struct siw_qp *qp = to_siw_qp(base_qp); local 999 struct siw_qp *qp = to_siw_qp(base_qp); local 1809 siw_qp_event(struct siw_qp *qp, enum ib_event_type etype) argument [all...] |
/linux-master/drivers/interconnect/qcom/ |
H A D | icc-rpmh.c | 25 struct qcom_icc_provider *qp; local 28 qp = to_qcom_provider(node->provider); 36 qcom_icc_bcm_voter_add(qp->voter, qn->bcms[i]); 88 struct qcom_icc_provider *qp; local 96 qp = to_qcom_provider(node->provider); 98 qcom_icc_bcm_voter_commit(qp->voter); 169 struct qcom_icc_provider *qp; local 181 qp = devm_kzalloc(dev, sizeof(*qp), GFP_KERNEL); 182 if (!qp) 258 struct qcom_icc_provider *qp = platform_get_drvdata(pdev); local [all...] |
/linux-master/drivers/scsi/bnx2i/ |
H A D | bnx2i_hwi.c | 153 cq_db = (struct bnx2i_5771x_cq_db *) ep->qp.cq_pgtbl_virt; 170 cq_index = ep->qp.cqe_exp_seq_sn + next_index - 1; 171 if (cq_index > ep->qp.cqe_size * 2) 172 cq_index -= ep->qp.cqe_size * 2; 195 if (!bnx2i_conn->ep->qp.rqe_left) 198 bnx2i_conn->ep->qp.rqe_left--; 199 memcpy(ptr, (u8 *) bnx2i_conn->ep->qp.rq_cons_qe, len); 200 if (bnx2i_conn->ep->qp.rq_cons_qe == bnx2i_conn->ep->qp.rq_last_qe) { 201 bnx2i_conn->ep->qp 1960 struct qp_info *qp; local [all...] |
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_mw.c | 50 static int rxe_check_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, argument 76 if (unlikely(qp->pd != to_rpd(mw->ibmw.pd))) { 78 "attempt to bind type 2 MW with qp with different PD\n"); 135 static void rxe_do_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, argument 159 rxe_get(qp); 160 mw->qp = qp; 164 int rxe_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe) argument 169 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 208 ret = rxe_check_bind_mw(qp, wq 224 rxe_check_invalidate_mw(struct rxe_qp *qp, struct rxe_mw *mw) argument 238 struct rxe_qp *qp; local 258 rxe_invalidate_mw(struct rxe_qp *qp, u32 rkey) argument 290 rxe_lookup_mw(struct rxe_qp *qp, int access, u32 rkey) argument 328 struct rxe_qp *qp = mw->qp; local [all...] |
H A D | rxe_av.c | 22 struct rxe_qp *qp; local 30 qp = obj; 31 rxe = to_rdev(qp->ibqp.device); 42 rxe_dbg_qp(qp, "invalid sgid index = %d\n", 54 rxe_dbg_qp(qp, "invalid network type for rdma_rxe = %d\n", 63 int rxe_av_chk_attr(struct rxe_qp *qp, struct rdma_ah_attr *attr) argument 65 return chk_attr(qp, attr, false); 139 if (!pkt || !pkt->qp) 142 if (qp_type(pkt->qp) == IB_QPT_RC || qp_type(pkt->qp) [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_main.c | 57 struct rvt_qp *qp; local 61 qp = rvt_lookup_qpn(ib_to_rvt(priv->device), &ibp->rvp, qpn); 62 if (!qp) { 67 rvt_get_qp(qp); 68 priv->qp = qp; 82 if (!priv->qp) 88 rvt_put_qp(priv->qp); 89 priv->qp = NULL; 111 struct rvt_qp *qp; local 141 struct rvt_qp *qp; local [all...] |
/linux-master/drivers/interconnect/ |
H A D | icc-clk.c | 81 struct icc_clk_provider *qp; local 91 qp = devm_kzalloc(dev, struct_size(qp, clocks, num_clocks), GFP_KERNEL); 92 if (!qp) 95 qp->num_clocks = num_clocks; 97 provider = &qp->provider; 109 qp->clocks[i].clk = data[i].clk; 118 node->data = &qp->clocks[i]; 157 struct icc_clk_provider *qp = container_of(provider, struct icc_clk_provider, provider); local 160 icc_provider_deregister(&qp [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.c | 110 if (unlikely(conn->qp.rq.pc - conn->qp.rq.cc >= conn->qp.rq.size)) { 115 ix = conn->qp.rq.pc & (conn->qp.rq.size - 1); 116 data = mlx5_wq_cyc_get_wqe(&conn->qp.wq.rq, ix); 121 conn->qp.rq.pc++; 122 conn->qp.rq.bufs[ix] = buf; 126 *conn->qp.wq.rq.db = cpu_to_be32(conn->qp [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_verbs.c | 216 * @qp: the QP the packet came on 223 int has_grh, void *data, u32 tlen, struct rvt_qp *qp) 227 spin_lock(&qp->r_lock); 230 if (!(ib_rvt_state_ops[qp->state] & RVT_PROCESS_RECV_OK)) { 235 switch (qp->ibqp.qp_type) { 242 qib_ud_rcv(ibp, hdr, has_grh, data, tlen, qp); 246 qib_rc_rcv(rcd, hdr, has_grh, data, tlen, qp); 250 qib_uc_rcv(ibp, hdr, has_grh, data, tlen, qp); 258 spin_unlock(&qp->r_lock); 279 struct rvt_qp *qp; local 222 qib_qp_rcv(struct qib_ctxtdata *rcd, struct ib_header *hdr, int has_grh, void *data, u32 tlen, struct rvt_qp *qp) argument 366 struct rvt_qp *qp = NULL; local 559 __get_txreq(struct qib_ibdev *dev, struct rvt_qp *qp) argument 591 get_txreq(struct qib_ibdev *dev, struct rvt_qp *qp) argument 616 struct rvt_qp *qp; local 669 struct rvt_qp *qp; local 715 struct rvt_qp *qp = tx->qp; local 746 wait_kmem(struct qib_ibdev *dev, struct rvt_qp *qp) argument 770 qib_verbs_send_dma(struct rvt_qp *qp, struct ib_header *hdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 874 no_bufs_available(struct rvt_qp *qp) argument 906 qib_verbs_send_pio(struct rvt_qp *qp, struct ib_header *ibhdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 1014 qib_verbs_send(struct rvt_qp *qp, struct ib_header *hdr, u32 hdrwords, struct rvt_sge_state *ss, u32 len) argument 1148 struct rvt_qp *qp; local 1683 _qib_schedule_send(struct rvt_qp *qp) argument 1700 qib_schedule_send(struct rvt_qp *qp) argument [all...] |
H A D | qib_verbs.h | 218 static inline int qib_send_ok(struct rvt_qp *qp) argument 220 return !(qp->s_flags & (RVT_S_BUSY | RVT_S_ANY_WAIT_IO)) && 221 (qp->s_hdrwords || (qp->s_flags & RVT_S_RESP_PENDING) || 222 !(qp->s_flags & RVT_S_ANY_WAIT_SEND)); 225 bool _qib_schedule_send(struct rvt_qp *qp); 226 bool qib_schedule_send(struct rvt_qp *qp); 272 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp); 273 void qib_qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp); 274 void qib_notify_qp_reset(struct rvt_qp *qp); [all...] |
/linux-master/drivers/scsi/qla2xxx/ |
H A D | qla_mid.c | 1084 int qla_create_buf_pool(struct scsi_qla_host *vha, struct qla_qpair *qp) argument 1088 qp->buf_pool.num_bufs = qp->req->length; 1090 sz = BITS_TO_LONGS(qp->req->length); 1091 qp->buf_pool.buf_map = kcalloc(sz, sizeof(long), GFP_KERNEL); 1092 if (!qp->buf_pool.buf_map) { 1097 sz = qp->req->length * sizeof(void *); 1098 qp->buf_pool.buf_array = kcalloc(qp->req->length, sizeof(void *), GFP_KERNEL); 1099 if (!qp 1118 qla_free_buf_pool(struct qla_qpair *qp) argument 1137 qla_get_buf(struct scsi_qla_host *vha, struct qla_qpair *qp, struct qla_buf_dsc *dsc) argument 1189 qla_trim_buf(struct qla_qpair *qp, u16 trim) argument 1221 __qla_adjust_buf(struct qla_qpair *qp) argument 1243 qla_put_buf(struct qla_qpair *qp, struct qla_buf_dsc *dsc) argument 1262 struct qla_qpair *qp; local [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_roce.h | 23 int qed_roce_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp); 26 struct qed_rdma_qp *qp, 30 struct qed_rdma_qp *qp,
|
/linux-master/drivers/net/ |
H A D | ntb_netdev.c | 76 struct ntb_transport_qp *qp; member in struct:ntb_netdev 89 ntb_transport_link_query(dev->qp)); 92 if (ntb_transport_link_query(dev->qp)) 99 static void ntb_netdev_rx_handler(struct ntb_transport_qp *qp, void *qp_data, argument 138 rc = ntb_transport_rx_enqueue(qp, skb, skb->data, ndev->mtu + ETH_HLEN); 147 struct ntb_transport_qp *qp, int size) 157 if (likely(ntb_transport_tx_free_entry(qp) < size)) { 167 struct ntb_transport_qp *qp, int size) 170 (ntb_transport_tx_free_entry(qp) >= size)) 173 return __ntb_netdev_maybe_stop_tx(ndev, qp, siz 146 __ntb_netdev_maybe_stop_tx(struct net_device *netdev, struct ntb_transport_qp *qp, int size) argument 166 ntb_netdev_maybe_stop_tx(struct net_device *ndev, struct ntb_transport_qp *qp, int size) argument 176 ntb_netdev_tx_handler(struct ntb_transport_qp *qp, void *qp_data, void *data, int len) argument [all...] |
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 1081 static int ocrdma_add_qpn_map(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument 1085 if (qp->id < OCRDMA_MAX_QP && dev->qp_tbl[qp->id] == NULL) { 1086 dev->qp_tbl[qp->id] = qp; 1092 static void ocrdma_del_qpn_map(struct ocrdma_dev *dev, struct ocrdma_qp *qp) argument 1094 dev->qp_tbl[qp->id] = NULL; 1105 pr_err("%s(%d) unsupported qp type=0x%x requested\n", 1171 static int ocrdma_copy_qp_uresp(struct ocrdma_qp *qp, argument 1178 struct ocrdma_pd *pd = qp 1230 ocrdma_set_qp_db(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ocrdma_pd *pd) argument 1250 ocrdma_alloc_wr_id_tbl(struct ocrdma_qp *qp) argument 1265 ocrdma_set_qp_init_params(struct ocrdma_qp *qp, struct ocrdma_pd *pd, struct ib_qp_init_attr *attrs) argument 1299 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local 1368 struct ocrdma_qp *qp; local 1389 struct ocrdma_qp *qp; local 1462 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local 1556 is_hw_sq_empty(struct ocrdma_qp *qp) argument 1561 is_hw_rq_empty(struct ocrdma_qp *qp) argument 1588 ocrdma_discard_cqes(struct ocrdma_qp *qp, struct ocrdma_cq *cq) argument 1650 ocrdma_del_flush_qp(struct ocrdma_qp *qp) argument 1672 struct ocrdma_qp *qp; local 1872 ocrdma_build_ud_hdr(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1918 ocrdma_build_inline_sges(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, struct ocrdma_sge *sge, const struct ib_send_wr *wr, u32 wqe_size) argument 1958 ocrdma_build_send(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1975 ocrdma_build_write(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1993 ocrdma_build_read(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 2022 ocrdma_build_reg(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_reg_wr *wr) argument 2078 ocrdma_ring_sq_db(struct ocrdma_qp *qp) argument 2089 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local 2190 ocrdma_ring_rq_db(struct ocrdma_qp *qp) argument 2223 struct ocrdma_qp *qp = get_ocrdma_qp(ibqp); local 2399 ocrdma_update_wc(struct ocrdma_qp *qp, struct ib_wc *ibwc, u32 wqe_idx) argument 2437 ocrdma_set_cqe_status_flushed(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe) argument 2469 ocrdma_update_err_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument 2491 ocrdma_update_err_rcqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument 2501 ocrdma_update_err_scqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp, int status) argument 2511 ocrdma_poll_err_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument 2555 ocrdma_poll_success_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled) argument 2581 ocrdma_poll_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument 2624 ocrdma_update_free_srq_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, struct ocrdma_qp *qp) argument 2644 ocrdma_poll_err_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop, int status) argument 2679 ocrdma_poll_success_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc) argument 2713 ocrdma_poll_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, struct ib_wc *ibwc, bool *polled, bool *stop) argument 2758 struct ocrdma_qp *qp = NULL; local 2811 ocrdma_add_err_cqe(struct ocrdma_cq *cq, int num_entries, struct ocrdma_qp *qp, struct ib_wc *ibwc) argument 2843 struct ocrdma_qp *qp; local [all...] |