Searched refs:qp (Results 126 - 150 of 379) sorted by relevance

1234567891011>>

/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_hw.h204 static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes) argument
206 struct h_epa epa = qp->epas.kernel;
211 static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes) argument
213 struct h_epa epa = qp->epas.kernel;
218 static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes) argument
220 struct h_epa epa = qp->epas.kernel;
225 static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes) argument
227 struct h_epa epa = qp->epas.kernel;
H A Dehea_qmr.h292 static inline struct ehea_rwqe *ehea_get_next_rwqe(struct ehea_qp *qp, argument
298 queue = &qp->hw_rqueue1;
300 queue = &qp->hw_rqueue2;
302 queue = &qp->hw_rqueue3;
325 static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index) argument
327 struct hw_queue *queue = &qp->hw_rqueue1;
338 static inline void ehea_inc_rq1(struct ehea_qp *qp) argument
340 hw_qeit_inc(&qp->hw_rqueue1);
372 int ehea_destroy_qp(struct ehea_qp *qp);
/linux-master/include/rdma/
H A Drdma_counter.h51 int rdma_counter_bind_qp_auto(struct ib_qp *qp, u32 port);
52 int rdma_counter_unbind_qp(struct ib_qp *qp, bool force);
H A Duverbs_std_types.h147 struct ib_qp *qp, struct ib_device *device,
155 if (qp) {
156 atomic_inc(&qp->usecnt);
157 ibflow->qp = qp;
146 ib_set_flow(struct ib_uobject *uobj, struct ib_flow *ibflow, struct ib_qp *qp, struct ib_device *device, struct ib_uflow_resources *uflow_res) argument
/linux-master/drivers/infiniband/hw/hfi1/
H A Dfault.h37 bool hfi1_dbg_should_fault_tx(struct rvt_qp *qp, u32 opcode);
54 static inline bool hfi1_dbg_should_fault_tx(struct rvt_qp *qp, argument
/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dib_verbs.h212 int bnxt_re_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *qp_init_attr,
214 int bnxt_re_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
216 int bnxt_re_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
218 int bnxt_re_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
219 int bnxt_re_post_send(struct ib_qp *qp, const struct ib_send_wr *send_wr,
221 int bnxt_re_post_recv(struct ib_qp *qp, const struct ib_recv_wr *recv_wr,
252 unsigned long bnxt_re_lock_cqs(struct bnxt_re_qp *qp);
253 void bnxt_re_unlock_cqs(struct bnxt_re_qp *qp, unsigned long flags);
H A Dqplib_fp.h530 int bnxt_qplib_create_qp1(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp);
531 int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp);
532 int bnxt_qplib_modify_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp);
533 int bnxt_qplib_query_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp);
534 int bnxt_qplib_destroy_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp);
535 void bnxt_qplib_clean_qp(struct bnxt_qplib_qp *qp);
537 struct bnxt_qplib_qp *qp);
538 void *bnxt_qplib_get_qp1_sq_buf(struct bnxt_qplib_qp *qp,
540 void *bnxt_qplib_get_qp1_rq_buf(struct bnxt_qplib_qp *qp,
542 u32 bnxt_qplib_get_rq_prod_index(struct bnxt_qplib_qp *qp);
[all...]
/linux-master/drivers/scsi/sym53c8xx_2/
H A Dsym_misc.h146 #define FOR_EACH_QUEUED_ELEMENT(head, qp) \
147 for (qp = (head)->flink; qp != (head); qp = qp->flink)
/linux-master/drivers/infiniband/core/
H A Dcore_priv.h226 int ib_security_modify_qp(struct ib_qp *qp,
231 int ib_create_qp_security(struct ib_qp *qp, struct ib_device *dev);
235 int ib_open_shared_qp_security(struct ib_qp *qp, struct ib_device *dev);
253 static inline int ib_security_modify_qp(struct ib_qp *qp, argument
258 return qp->device->ops.modify_qp(qp->real_qp,
264 static inline int ib_create_qp_security(struct ib_qp *qp, argument
282 static inline int ib_open_shared_qp_security(struct ib_qp *qp, argument
324 void ib_qp_usecnt_inc(struct ib_qp *qp);
325 void ib_qp_usecnt_dec(struct ib_qp *qp);
[all...]
H A Drestrack.c239 struct ib_qp *qp = container_of(res, struct ib_qp, res); local
241 WARN_ONCE(qp->qp_num >> 24 || qp->port >> 8,
242 "QP number 0x%0X and port 0x%0X", qp->qp_num,
243 qp->port);
244 res->id = qp->qp_num;
245 if (qp->qp_type == IB_QPT_SMI || qp->qp_type == IB_QPT_GSI)
246 res->id |= qp->port << 24;
/linux-master/drivers/scsi/lpfc/
H A Dlpfc_debugfs.c406 struct lpfc_sli4_hdw_queue *qp; local
414 qp = &phba->sli4_hba.hdwq[lpfc_debugfs_last_xripool];
417 spin_lock_irqsave(&qp->abts_io_buf_list_lock, iflag);
418 spin_lock(&qp->io_buf_list_get_lock);
419 spin_lock(&qp->io_buf_list_put_lock);
420 out = qp->total_io_bufs - (qp->get_io_bufs + qp->put_io_bufs +
421 qp->abts_scsi_io_bufs + qp
460 struct lpfc_sli4_hdw_queue *qp; local
596 struct lpfc_sli4_hdw_queue *qp; local
2154 struct lpfc_sli4_hdw_queue *qp; local
2612 struct lpfc_sli4_hdw_queue *qp; local
3859 __lpfc_idiag_print_wq(struct lpfc_queue *qp, char *wqtype, char *pbuffer, int len) argument
3886 struct lpfc_queue *qp; local
3901 __lpfc_idiag_print_cq(struct lpfc_queue *qp, char *cqtype, char *pbuffer, int len) argument
3928 __lpfc_idiag_print_rqpair(struct lpfc_queue *qp, struct lpfc_queue *datqp, char *rqtype, char *pbuffer, int len) argument
3959 struct lpfc_queue *qp; local
4002 __lpfc_idiag_print_eq(struct lpfc_queue *qp, char *eqtype, char *pbuffer, int len) argument
4051 struct lpfc_queue *qp = NULL; local
4337 struct lpfc_queue *pque, *qp; local
[all...]
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_io.c113 struct hinic_qp *qp; local
129 qp = &func_to_io->qps[i];
131 hinic_sq_prepare_ctxt(&sq_ctxt[i], &qp->sq,
132 base_qpn + qp->q_id);
157 struct hinic_qp *qp; local
173 qp = &func_to_io->qps[i];
175 hinic_rq_prepare_ctxt(&rq_ctxt[i], &qp->rq,
176 base_qpn + qp->q_id);
194 * write_qp_ctxts - write the qp ctxt to HW
196 * @base_qpn: first qp numbe
271 init_qp(struct hinic_func_to_io *func_to_io, struct hinic_qp *qp, int q_id, struct msix_entry *sq_msix_entry, struct msix_entry *rq_msix_entry) argument
347 destroy_qp(struct hinic_func_to_io *func_to_io, struct hinic_qp *qp) argument
[all...]
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_mcg.c43 __be32 qp[MTHCA_QP_PER_MGM]; member in struct:mthca_mgm
165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) {
170 } else if (!(mgm->qp[i] & cpu_to_be32(1 << 31))) {
171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31));
242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31)))
244 if (!(mgm->qp[i] & cpu_to_be32(1 << 31)))
254 mgm->qp[loc] = mgm->qp[i - 1];
255 mgm->qp[i - 1] = 0;
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_eq.c34 struct erdma_qp *qp; local
67 qp = find_qp_by_qpn(dev, qpn);
68 if (!qp)
71 event.device = qp->ibqp.device;
72 event.element.qp = &qp->ibqp;
74 if (qp->ibqp.event_handler)
75 qp->ibqp.event_handler(&event,
76 qp->ibqp.qp_context);
H A Derdma_verbs.h271 #define QP_ID(qp) ((qp)->ibqp.qp_num)
283 void erdma_qp_get(struct erdma_qp *qp);
284 void erdma_qp_put(struct erdma_qp *qp);
285 int erdma_modify_qp_internal(struct erdma_qp *qp, struct erdma_qp_attrs *attrs,
287 void erdma_qp_llp_close(struct erdma_qp *qp);
288 void erdma_qp_cm_drop(struct erdma_qp *qp);
305 static inline struct erdma_qp *to_eqp(struct ib_qp *qp) argument
307 return container_of(qp, struct erdma_qp, ibqp);
/linux-master/drivers/infiniband/hw/irdma/
H A Dutils.c794 * irdma_get_qp - get qp address
796 * @qpn: qp number
860 * @qp: hardware control qp
863 int irdma_cqp_qp_suspend_resume(struct irdma_sc_qp *qp, u8 op) argument
865 struct irdma_sc_dev *dev = qp->dev;
879 cqp_info->in.u.suspend_resume.qp = qp;
889 * irdma_term_modify_qp - modify qp for term message
890 * @qp
895 irdma_term_modify_qp(struct irdma_sc_qp *qp, u8 next_state, u8 term, u8 term_len) argument
909 irdma_terminate_done(struct irdma_sc_qp *qp, int timeout_occurred) argument
937 struct irdma_sc_qp *qp = &iwqp->sc_qp; local
947 irdma_terminate_start_timer(struct irdma_sc_qp *qp) argument
963 irdma_terminate_del_timer(struct irdma_sc_qp *qp) argument
1076 irdma_cqp_qp_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument
1110 irdma_dealloc_push_page(struct irdma_pci_f *rf, struct irdma_sc_qp *qp) argument
1305 irdma_cqp_qp_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument
1336 irdma_ieq_mpa_crc_ae(struct irdma_sc_dev *dev, struct irdma_sc_qp *qp) argument
1447 irdma_send_ieq_ack(struct irdma_sc_qp *qp) argument
1464 irdma_puda_ieq_get_ah_info(struct irdma_sc_qp *qp, struct irdma_ah_info *ah_info) argument
1998 struct irdma_sc_qp *qp = cqp_request->param; local
2339 struct irdma_qp *qp = sc_qp->qp_uk.back_qp; local
2427 irdma_set_cpi_common_values(struct irdma_cq_poll_info *cpi, struct irdma_qp_uk *qp, u32 qp_num) argument
2448 struct irdma_qp_uk *qp = &iwqp->sc_qp.qp_uk; local
[all...]
H A Dhw.c134 static void irdma_set_flush_fields(struct irdma_sc_qp *qp, argument
137 qp->sq_flush_code = info->sq;
138 qp->rq_flush_code = info->rq;
139 qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC;
152 qp->flush_code = FLUSH_PROT_ERR;
153 qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
157 qp->flush_code = FLUSH_LOC_QP_OP_ERR;
158 qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC;
165 qp->flush_code = FLUSH_LOC_LEN_ERR;
166 qp
223 struct irdma_sc_qp *qp = NULL; local
2548 struct irdma_sc_qp *qp; local
2585 irdma_hw_flush_wqes(struct irdma_pci_f *rf, struct irdma_sc_qp *qp, struct irdma_qp_flush_info *info, bool wait) argument
2694 irdma_gen_ae(struct irdma_pci_f *rf, struct irdma_sc_qp *qp, struct irdma_gen_ae_info *info, bool wait) argument
[all...]
/linux-master/drivers/infiniband/hw/mlx5/
H A Dwr.h95 int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg,
99 void mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl,
102 void mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq,
H A DMakefile17 qp.o \
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Dmcg.c38 #include <linux/mlx4/qp.h>
173 /* If the given qpn is also a promisc qp,
218 mgm->qp[members_count++] = cpu_to_be32(pqp->qpn & MGM_QPN_MASK);
273 return 0; /* qp is already duplicated */
276 /* add the qp as a duplicate on this index */
301 /* if qp is not promisc, it cannot be duplicated */
305 /* The qp is promisc qp so it is a duplicate on this index
355 u32 qpn = be32_to_cpu(mgm->qp[i]) & MGM_QPN_MASK;
357 /* the qp i
1104 mlx4_qp_attach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot, enum mlx4_steer_type steer) argument
1214 mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, enum mlx4_steer_type steer) argument
1335 mlx4_QP_ATTACH(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 attach, u8 block_loopback, enum mlx4_protocol prot) argument
1367 mlx4_trans_to_dmfs_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) argument
1407 mlx4_multicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) argument
1438 mlx4_multicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, u64 reg_id) argument
1529 mlx4_unicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot) argument
1545 mlx4_unicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot) argument
[all...]
H A DMakefile5 main.o mcg.o mr.o pd.o port.o profile.o qp.o reset.o sense.o \
/linux-master/drivers/infiniband/hw/efa/
H A Defa_verbs.c309 struct efa_qp *qp = to_eqp(ibqp); local
326 params.qp_handle = qp->qp_handle;
338 qp_attr->cap.max_send_wr = qp->max_send_wr;
339 qp_attr->cap.max_recv_wr = qp->max_recv_wr;
340 qp_attr->cap.max_send_sge = qp->max_send_sge;
341 qp_attr->cap.max_recv_sge = qp->max_recv_sge;
342 qp_attr->cap.max_inline_data = qp->max_inline_data;
443 static void efa_qp_user_mmap_entries_remove(struct efa_qp *qp) argument
445 rdma_user_mmap_entry_remove(qp->rq_mmap_entry);
446 rdma_user_mmap_entry_remove(qp
454 struct efa_qp *qp = to_eqp(ibqp); local
502 qp_mmap_entries_setup(struct efa_qp *qp, struct efa_dev *dev, struct efa_ucontext *ucontext, struct efa_com_create_qp_params *params, struct efa_ibv_create_qp_resp *resp) argument
640 struct efa_qp *qp = to_eqp(ibqp); local
886 efa_modify_qp_validate(struct efa_dev *dev, struct efa_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) argument
935 struct efa_qp *qp = to_eqp(ibqp); local
[all...]
/linux-master/drivers/vfio/pci/mlx5/
H A Dcmd.c1166 struct mlx5_vhca_qp *qp; local
1174 qp = kzalloc(sizeof(*qp), GFP_KERNEL_ACCOUNT);
1175 if (!qp)
1178 err = mlx5_db_alloc_node(mdev, &qp->db, mdev->priv.numa_node);
1183 qp->rq.wqe_cnt = roundup_pow_of_two(max_recv_wr);
1185 log_rq_sz = ilog2(qp->rq.wqe_cnt);
1188 &qp->buf, mdev->priv.numa_node);
1191 mlx5_init_fbc(qp->buf.frags, log_rq_stride, log_rq_sz, &qp
1247 mlx5vf_post_recv(struct mlx5_vhca_qp *qp) argument
1264 mlx5vf_activate_qp(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 remote_qpn, bool host_qp) argument
1323 mlx5vf_destroy_qp(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp) argument
1426 mlx5vf_free_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp) argument
1436 mlx5vf_alloc_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 pdn, u64 rq_size) argument
1617 set_report_output(u32 size, int index, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty) argument
1644 mlx5vf_rq_cqe(struct mlx5_vhca_qp *qp, struct mlx5_cqe64 *cqe, struct iova_bitmap *dirty, int *tracker_status) argument
1685 mlx5vf_cq_poll_one(struct mlx5_vhca_cq *cq, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty, int *tracker_status) argument
[all...]
/linux-master/drivers/infiniband/ulp/ipoib/
H A Dipoib_cm.c129 ret = ib_post_recv(rx->qp, wr, NULL);
227 if (ib_post_send(p->qp, &ipoib_cm_rx_drain_wr, NULL))
274 struct ib_cm_id *cm_id, struct ib_qp *qp,
287 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask);
299 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask);
319 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask);
421 struct ib_qp *qp,
429 data.qpn = cpu_to_be32(priv->qp->qp_num);
437 rep.qp_num = qp->qp_num;
462 p->qp
273 ipoib_cm_modify_rx_qp(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, unsigned int psn) argument
420 ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, const struct ib_cm_req_event_param *req, unsigned int psn) argument
1082 ipoib_cm_send_req(struct net_device *dev, struct ib_cm_id *id, struct ib_qp *qp, u32 qpn, struct sa_path_rec *pathrec) argument
1119 ipoib_cm_modify_tx_init(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp) argument
[all...]
/linux-master/drivers/infiniband/hw/qib/
H A Dqib_sdma.c492 struct qib_qp_priv *priv = tx->qp->priv;
517 struct rvt_qp *qp; local
605 priv = tx->qp->priv;
623 qp = tx->qp;
624 priv = qp->priv;
626 spin_lock(&qp->r_lock);
627 spin_lock(&qp->s_lock);
628 if (qp->ibqp.qp_type == IB_QPT_RC) {
630 if (ib_rvt_state_ops[qp
[all...]

Completed in 394 milliseconds

1234567891011>>