/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_hw.h | 204 static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes) argument 206 struct h_epa epa = qp->epas.kernel; 211 static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes) argument 213 struct h_epa epa = qp->epas.kernel; 218 static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes) argument 220 struct h_epa epa = qp->epas.kernel; 225 static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes) argument 227 struct h_epa epa = qp->epas.kernel;
|
H A D | ehea_qmr.h | 292 static inline struct ehea_rwqe *ehea_get_next_rwqe(struct ehea_qp *qp, argument 298 queue = &qp->hw_rqueue1; 300 queue = &qp->hw_rqueue2; 302 queue = &qp->hw_rqueue3; 325 static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index) argument 327 struct hw_queue *queue = &qp->hw_rqueue1; 338 static inline void ehea_inc_rq1(struct ehea_qp *qp) argument 340 hw_qeit_inc(&qp->hw_rqueue1); 372 int ehea_destroy_qp(struct ehea_qp *qp);
|
/linux-master/include/rdma/ |
H A D | rdma_counter.h | 51 int rdma_counter_bind_qp_auto(struct ib_qp *qp, u32 port); 52 int rdma_counter_unbind_qp(struct ib_qp *qp, bool force);
|
H A D | uverbs_std_types.h | 147 struct ib_qp *qp, struct ib_device *device, 155 if (qp) { 156 atomic_inc(&qp->usecnt); 157 ibflow->qp = qp; 146 ib_set_flow(struct ib_uobject *uobj, struct ib_flow *ibflow, struct ib_qp *qp, struct ib_device *device, struct ib_uflow_resources *uflow_res) argument
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | fault.h | 37 bool hfi1_dbg_should_fault_tx(struct rvt_qp *qp, u32 opcode); 54 static inline bool hfi1_dbg_should_fault_tx(struct rvt_qp *qp, argument
|
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.h | 212 int bnxt_re_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *qp_init_attr, 214 int bnxt_re_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 216 int bnxt_re_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 218 int bnxt_re_destroy_qp(struct ib_qp *qp, struct ib_udata *udata); 219 int bnxt_re_post_send(struct ib_qp *qp, const struct ib_send_wr *send_wr, 221 int bnxt_re_post_recv(struct ib_qp *qp, const struct ib_recv_wr *recv_wr, 252 unsigned long bnxt_re_lock_cqs(struct bnxt_re_qp *qp); 253 void bnxt_re_unlock_cqs(struct bnxt_re_qp *qp, unsigned long flags);
|
H A D | qplib_fp.h | 530 int bnxt_qplib_create_qp1(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp); 531 int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp); 532 int bnxt_qplib_modify_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp); 533 int bnxt_qplib_query_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp); 534 int bnxt_qplib_destroy_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp); 535 void bnxt_qplib_clean_qp(struct bnxt_qplib_qp *qp); 537 struct bnxt_qplib_qp *qp); 538 void *bnxt_qplib_get_qp1_sq_buf(struct bnxt_qplib_qp *qp, 540 void *bnxt_qplib_get_qp1_rq_buf(struct bnxt_qplib_qp *qp, 542 u32 bnxt_qplib_get_rq_prod_index(struct bnxt_qplib_qp *qp); [all...] |
/linux-master/drivers/scsi/sym53c8xx_2/ |
H A D | sym_misc.h | 146 #define FOR_EACH_QUEUED_ELEMENT(head, qp) \ 147 for (qp = (head)->flink; qp != (head); qp = qp->flink)
|
/linux-master/drivers/infiniband/core/ |
H A D | core_priv.h | 226 int ib_security_modify_qp(struct ib_qp *qp, 231 int ib_create_qp_security(struct ib_qp *qp, struct ib_device *dev); 235 int ib_open_shared_qp_security(struct ib_qp *qp, struct ib_device *dev); 253 static inline int ib_security_modify_qp(struct ib_qp *qp, argument 258 return qp->device->ops.modify_qp(qp->real_qp, 264 static inline int ib_create_qp_security(struct ib_qp *qp, argument 282 static inline int ib_open_shared_qp_security(struct ib_qp *qp, argument 324 void ib_qp_usecnt_inc(struct ib_qp *qp); 325 void ib_qp_usecnt_dec(struct ib_qp *qp); [all...] |
H A D | restrack.c | 239 struct ib_qp *qp = container_of(res, struct ib_qp, res); local 241 WARN_ONCE(qp->qp_num >> 24 || qp->port >> 8, 242 "QP number 0x%0X and port 0x%0X", qp->qp_num, 243 qp->port); 244 res->id = qp->qp_num; 245 if (qp->qp_type == IB_QPT_SMI || qp->qp_type == IB_QPT_GSI) 246 res->id |= qp->port << 24;
|
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.c | 406 struct lpfc_sli4_hdw_queue *qp; local 414 qp = &phba->sli4_hba.hdwq[lpfc_debugfs_last_xripool]; 417 spin_lock_irqsave(&qp->abts_io_buf_list_lock, iflag); 418 spin_lock(&qp->io_buf_list_get_lock); 419 spin_lock(&qp->io_buf_list_put_lock); 420 out = qp->total_io_bufs - (qp->get_io_bufs + qp->put_io_bufs + 421 qp->abts_scsi_io_bufs + qp 460 struct lpfc_sli4_hdw_queue *qp; local 596 struct lpfc_sli4_hdw_queue *qp; local 2154 struct lpfc_sli4_hdw_queue *qp; local 2612 struct lpfc_sli4_hdw_queue *qp; local 3859 __lpfc_idiag_print_wq(struct lpfc_queue *qp, char *wqtype, char *pbuffer, int len) argument 3886 struct lpfc_queue *qp; local 3901 __lpfc_idiag_print_cq(struct lpfc_queue *qp, char *cqtype, char *pbuffer, int len) argument 3928 __lpfc_idiag_print_rqpair(struct lpfc_queue *qp, struct lpfc_queue *datqp, char *rqtype, char *pbuffer, int len) argument 3959 struct lpfc_queue *qp; local 4002 __lpfc_idiag_print_eq(struct lpfc_queue *qp, char *eqtype, char *pbuffer, int len) argument 4051 struct lpfc_queue *qp = NULL; local 4337 struct lpfc_queue *pque, *qp; local [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_io.c | 113 struct hinic_qp *qp; local 129 qp = &func_to_io->qps[i]; 131 hinic_sq_prepare_ctxt(&sq_ctxt[i], &qp->sq, 132 base_qpn + qp->q_id); 157 struct hinic_qp *qp; local 173 qp = &func_to_io->qps[i]; 175 hinic_rq_prepare_ctxt(&rq_ctxt[i], &qp->rq, 176 base_qpn + qp->q_id); 194 * write_qp_ctxts - write the qp ctxt to HW 196 * @base_qpn: first qp numbe 271 init_qp(struct hinic_func_to_io *func_to_io, struct hinic_qp *qp, int q_id, struct msix_entry *sq_msix_entry, struct msix_entry *rq_msix_entry) argument 347 destroy_qp(struct hinic_func_to_io *func_to_io, struct hinic_qp *qp) argument [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_mcg.c | 43 __be32 qp[MTHCA_QP_PER_MGM]; member in struct:mthca_mgm 165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { 170 } else if (!(mgm->qp[i] & cpu_to_be32(1 << 31))) { 171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); 242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) 244 if (!(mgm->qp[i] & cpu_to_be32(1 << 31))) 254 mgm->qp[loc] = mgm->qp[i - 1]; 255 mgm->qp[i - 1] = 0;
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_eq.c | 34 struct erdma_qp *qp; local 67 qp = find_qp_by_qpn(dev, qpn); 68 if (!qp) 71 event.device = qp->ibqp.device; 72 event.element.qp = &qp->ibqp; 74 if (qp->ibqp.event_handler) 75 qp->ibqp.event_handler(&event, 76 qp->ibqp.qp_context);
|
H A D | erdma_verbs.h | 271 #define QP_ID(qp) ((qp)->ibqp.qp_num) 283 void erdma_qp_get(struct erdma_qp *qp); 284 void erdma_qp_put(struct erdma_qp *qp); 285 int erdma_modify_qp_internal(struct erdma_qp *qp, struct erdma_qp_attrs *attrs, 287 void erdma_qp_llp_close(struct erdma_qp *qp); 288 void erdma_qp_cm_drop(struct erdma_qp *qp); 305 static inline struct erdma_qp *to_eqp(struct ib_qp *qp) argument 307 return container_of(qp, struct erdma_qp, ibqp);
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | utils.c | 794 * irdma_get_qp - get qp address 796 * @qpn: qp number 860 * @qp: hardware control qp 863 int irdma_cqp_qp_suspend_resume(struct irdma_sc_qp *qp, u8 op) argument 865 struct irdma_sc_dev *dev = qp->dev; 879 cqp_info->in.u.suspend_resume.qp = qp; 889 * irdma_term_modify_qp - modify qp for term message 890 * @qp 895 irdma_term_modify_qp(struct irdma_sc_qp *qp, u8 next_state, u8 term, u8 term_len) argument 909 irdma_terminate_done(struct irdma_sc_qp *qp, int timeout_occurred) argument 937 struct irdma_sc_qp *qp = &iwqp->sc_qp; local 947 irdma_terminate_start_timer(struct irdma_sc_qp *qp) argument 963 irdma_terminate_del_timer(struct irdma_sc_qp *qp) argument 1076 irdma_cqp_qp_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument 1110 irdma_dealloc_push_page(struct irdma_pci_f *rf, struct irdma_sc_qp *qp) argument 1305 irdma_cqp_qp_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument 1336 irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument 1447 irdma_send_ieq_ack(struct irdma_sc_qp *qp) argument 1464 irdma_puda_ieq_get_ah_info(struct irdma_sc_qp *qp, struct irdma_ah_info *ah_info) argument 1998 struct irdma_sc_qp *qp = cqp_request->param; local 2339 struct irdma_qp *qp = sc_qp->qp_uk.back_qp; local 2427 irdma_set_cpi_common_values(struct irdma_cq_poll_info *cpi, struct irdma_qp_uk *qp, u32 qp_num) argument 2448 struct irdma_qp_uk *qp = &iwqp->sc_qp.qp_uk; local [all...] |
H A D | hw.c | 134 static void irdma_set_flush_fields(struct irdma_sc_qp *qp, argument 137 qp->sq_flush_code = info->sq; 138 qp->rq_flush_code = info->rq; 139 qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC; 152 qp->flush_code = FLUSH_PROT_ERR; 153 qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR; 157 qp->flush_code = FLUSH_LOC_QP_OP_ERR; 158 qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC; 165 qp->flush_code = FLUSH_LOC_LEN_ERR; 166 qp 223 struct irdma_sc_qp *qp = NULL; local 2548 struct irdma_sc_qp *qp; local 2585 irdma_hw_flush_wqes(struct irdma_pci_f *rf, struct irdma_sc_qp *qp, struct irdma_qp_flush_info *info, bool wait) argument 2694 irdma_gen_ae(struct irdma_pci_f *rf, struct irdma_sc_qp *qp, struct irdma_gen_ae_info *info, bool wait) argument [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | wr.h | 95 int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, 99 void mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl, 102 void mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq,
|
H A D | Makefile | 17 qp.o \
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mcg.c | 38 #include <linux/mlx4/qp.h> 173 /* If the given qpn is also a promisc qp, 218 mgm->qp[members_count++] = cpu_to_be32(pqp->qpn & MGM_QPN_MASK); 273 return 0; /* qp is already duplicated */ 276 /* add the qp as a duplicate on this index */ 301 /* if qp is not promisc, it cannot be duplicated */ 305 /* The qp is promisc qp so it is a duplicate on this index 355 u32 qpn = be32_to_cpu(mgm->qp[i]) & MGM_QPN_MASK; 357 /* the qp i 1104 mlx4_qp_attach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot, enum mlx4_steer_type steer) argument 1214 mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, enum mlx4_steer_type steer) argument 1335 mlx4_QP_ATTACH(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 attach, u8 block_loopback, enum mlx4_protocol prot) argument 1367 mlx4_trans_to_dmfs_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) argument 1407 mlx4_multicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) argument 1438 mlx4_multicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, u64 reg_id) argument 1529 mlx4_unicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot) argument 1545 mlx4_unicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot) argument [all...] |
H A D | Makefile | 5 main.o mcg.o mr.o pd.o port.o profile.o qp.o reset.o sense.o \
|
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 309 struct efa_qp *qp = to_eqp(ibqp); local 326 params.qp_handle = qp->qp_handle; 338 qp_attr->cap.max_send_wr = qp->max_send_wr; 339 qp_attr->cap.max_recv_wr = qp->max_recv_wr; 340 qp_attr->cap.max_send_sge = qp->max_send_sge; 341 qp_attr->cap.max_recv_sge = qp->max_recv_sge; 342 qp_attr->cap.max_inline_data = qp->max_inline_data; 443 static void efa_qp_user_mmap_entries_remove(struct efa_qp *qp) argument 445 rdma_user_mmap_entry_remove(qp->rq_mmap_entry); 446 rdma_user_mmap_entry_remove(qp 454 struct efa_qp *qp = to_eqp(ibqp); local 502 qp_mmap_entries_setup(struct efa_qp *qp, struct efa_dev *dev, struct efa_ucontext *ucontext, struct efa_com_create_qp_params *params, struct efa_ibv_create_qp_resp *resp) argument 640 struct efa_qp *qp = to_eqp(ibqp); local 886 efa_modify_qp_validate(struct efa_dev *dev, struct efa_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) argument 935 struct efa_qp *qp = to_eqp(ibqp); local [all...] |
/linux-master/drivers/vfio/pci/mlx5/ |
H A D | cmd.c | 1166 struct mlx5_vhca_qp *qp; local 1174 qp = kzalloc(sizeof(*qp), GFP_KERNEL_ACCOUNT); 1175 if (!qp) 1178 err = mlx5_db_alloc_node(mdev, &qp->db, mdev->priv.numa_node); 1183 qp->rq.wqe_cnt = roundup_pow_of_two(max_recv_wr); 1185 log_rq_sz = ilog2(qp->rq.wqe_cnt); 1188 &qp->buf, mdev->priv.numa_node); 1191 mlx5_init_fbc(qp->buf.frags, log_rq_stride, log_rq_sz, &qp 1247 mlx5vf_post_recv(struct mlx5_vhca_qp *qp) argument 1264 mlx5vf_activate_qp(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 remote_qpn, bool host_qp) argument 1323 mlx5vf_destroy_qp(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp) argument 1426 mlx5vf_free_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp) argument 1436 mlx5vf_alloc_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 pdn, u64 rq_size) argument 1617 set_report_output(u32 size, int index, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty) argument 1644 mlx5vf_rq_cqe(struct mlx5_vhca_qp *qp, struct mlx5_cqe64 *cqe, struct iova_bitmap *dirty, int *tracker_status) argument 1685 mlx5vf_cq_poll_one(struct mlx5_vhca_cq *cq, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty, int *tracker_status) argument [all...] |
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 129 ret = ib_post_recv(rx->qp, wr, NULL); 227 if (ib_post_send(p->qp, &ipoib_cm_rx_drain_wr, NULL)) 274 struct ib_cm_id *cm_id, struct ib_qp *qp, 287 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); 299 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); 319 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); 421 struct ib_qp *qp, 429 data.qpn = cpu_to_be32(priv->qp->qp_num); 437 rep.qp_num = qp->qp_num; 462 p->qp 273 ipoib_cm_modify_rx_qp(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, unsigned int psn) argument 420 ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, const struct ib_cm_req_event_param *req, unsigned int psn) argument 1082 ipoib_cm_send_req(struct net_device *dev, struct ib_cm_id *id, struct ib_qp *qp, u32 qpn, struct sa_path_rec *pathrec) argument 1119 ipoib_cm_modify_tx_init(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp) argument [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_sdma.c | 492 struct qib_qp_priv *priv = tx->qp->priv; 517 struct rvt_qp *qp; local 605 priv = tx->qp->priv; 623 qp = tx->qp; 624 priv = qp->priv; 626 spin_lock(&qp->r_lock); 627 spin_lock(&qp->s_lock); 628 if (qp->ibqp.qp_type == IB_QPT_RC) { 630 if (ib_rvt_state_ops[qp [all...] |