/linux-master/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs.c | 93 return ib_post_recv(con->qp, &wr, NULL); 105 return ib_post_recv(con->qp, &wr, NULL); 109 static int rtrs_post_send(struct ib_qp *qp, struct ib_send_wr *head, argument 125 return ib_post_send(qp, head, NULL); 150 return rtrs_post_send(con->qp, head, &wr, NULL); 183 return rtrs_post_send(con->qp, head, &wr.wr, tail); 207 return rtrs_post_send(con->qp, head, &wr.wr, NULL); 279 con->qp = cm_id->qp; 320 if (con->qp) { [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 500 struct hinic_qp *qp; local 503 qp = container_of(txq->sq, struct hinic_qp, sq); 514 netif_stop_subqueue(netdev, qp->q_id); 518 netif_wake_subqueue(nic_dev->netdev, qp->q_id); 561 struct hinic_qp *qp; local 564 qp = container_of(txq->sq, struct hinic_qp, sq); 595 netif_stop_subqueue(netdev, qp->q_id); 602 netif_wake_subqueue(nic_dev->netdev, qp->q_id); 700 struct hinic_qp *qp = container_of(txq->sq, struct hinic_qp, sq); local 744 if (__netif_subqueue_stopped(nic_dev->netdev, qp 805 struct hinic_qp *qp; local 861 struct hinic_qp *qp = container_of(sq, struct hinic_qp, sq); local [all...] |
H A D | hinic_rx.c | 356 struct hinic_qp *qp = container_of(rxq->rq, struct hinic_qp, rq); local 421 skb_record_rx_queue(skb, qp->q_id); 520 struct hinic_qp *qp; local 523 qp = container_of(rq, struct hinic_qp, rq); 532 intr_coal = &nic_dev->rx_intr_coalesce[qp->q_id]; 549 cpumask_set_cpu(qp->q_id % num_online_cpus(), &rq->affinity_mask); 583 struct hinic_qp *qp = container_of(rq, struct hinic_qp, rq); local 594 "%s_rxq%d", netdev->name, qp->q_id);
|
/linux-master/drivers/infiniband/core/ |
H A D | uverbs_std_types.c | 58 struct ib_qp *qp = flow->qp; local 63 if (qp) 64 atomic_dec(&qp->usecnt);
|
H A D | nldev.c | 411 [RDMA_RESTRACK_QP] = "qp", 481 struct ib_qp *qp) 487 ret = ib_query_qp(qp, &qp_attr, 0, &qp_init_attr); 491 if (qp->qp_type == IB_QPT_RC || qp->qp_type == IB_QPT_UC) { 503 if (qp->qp_type == IB_QPT_RC || qp->qp_type == IB_QPT_UC || 504 qp->qp_type == IB_QPT_XRC_INI || qp->qp_type == IB_QPT_XRC_TGT) { 509 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_TYPE, qp 478 fill_res_qp_entry_query(struct sk_buff *msg, struct rdma_restrack_entry *res, struct ib_device *dev, struct ib_qp *qp) argument 524 struct ib_qp *qp = container_of(res, struct ib_qp, res); local 553 struct ib_qp *qp = container_of(res, struct ib_qp, res); local 783 struct ib_qp *qp = NULL; local 927 struct ib_qp *qp = NULL; local [all...] |
/linux-master/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.h | 46 * The qp group struct represents all the hw resources needed to present a ib_qp 92 int usnic_ib_qp_grp_create(struct usnic_ib_qp_grp *qp,
|
H A D | usnic_ib_verbs.c | 171 find_free_vf_and_create_qp_grp(struct ib_qp *qp, argument 175 struct usnic_ib_dev *us_ibdev = to_usdev(qp->device); 176 struct usnic_ib_pd *pd = to_upd(qp->pd); 179 struct usnic_ib_qp_grp *qp_grp = to_uqp_grp(qp); 236 usnic_info("No free qp grp found on %s\n", 385 int usnic_ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, argument 398 qp_grp = to_uqp_grp(qp); 527 int usnic_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata) argument 534 qp_grp = to_uqp_grp(qp); 538 usnic_err("Failed to move qp gr [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | aso.h | 7 #include <linux/mlx5/qp.h>
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | btf_dump_test_case_padding.c | 180 void *qp; member in struct:ib_wc
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | Makefile | 38 qp.o \
|
/linux-master/include/linux/mlx4/ |
H A D | qp.h | 481 int sqd_event, struct mlx4_qp *qp); 483 int mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, 488 struct mlx4_qp *qp, enum mlx4_qp_state *qp_state); 495 void mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp); 507 void mlx4_put_qp(struct mlx4_qp *qp);
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_verbs.h | 339 struct rxe_qp *qp; /* Type 2 only */ member in struct:rxe_mw 360 struct rxe_qp *qp; member in struct:rxe_mca 439 static inline struct rxe_qp *to_rqp(struct ib_qp *qp) argument 441 return qp ? container_of(qp, struct rxe_qp, ibqp) : NULL;
|
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_main.c | 197 arr[i++].fwh = pr->qp->fw_handle; 419 ehea_update_rq1a(pr->qp, adder); 439 ehea_update_rq1a(pr->qp, i - 1); 447 struct ehea_qp *qp = pr->qp; local 489 rwqe = ehea_get_next_rwqe(qp, rq_nr); 509 ehea_update_rq2a(pr->qp, adder); 511 ehea_update_rq3a(pr->qp, adder); 643 pr->qp->init_attr.qp_nr); 658 struct ehea_qp *qp local 926 struct ehea_qp *qp; local 2165 ehea_activate_qp(struct ehea_adapter *adapter, struct ehea_qp *qp) argument 2477 struct ehea_qp qp = *orig_qp; local 2529 struct ehea_qp *qp = pr->qp; local 2580 struct ehea_qp qp = *orig_qp; local 2629 struct ehea_qp *qp = pr->qp; local [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 39 #include "qp.h" 167 struct mlx5_ib_qp *qp) 169 enum rdma_link_layer ll = rdma_port_get_link_layer(qp->ibqp.device, 1); 170 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); 178 if (qp->ibqp.srq || qp->ibqp.xrcd) { 181 if (qp->ibqp.xrcd) { 186 srq = to_msrq(qp->ibqp.srq); 196 wq = &qp->rq; 230 if (is_qp1(qp 166 handle_responder(struct ib_wc *wc, struct mlx5_cqe64 *cqe, struct mlx5_ib_qp *qp) argument 341 handle_atomics(struct mlx5_ib_qp *qp, struct mlx5_cqe64 *cqe64, u16 tail, u16 head) argument 394 sw_comp(struct mlx5_ib_qp *qp, int num_entries, struct ib_wc *wc, int *npolled, bool is_send) argument 430 struct mlx5_ib_qp *qp; local [all...] |
/linux-master/mm/ |
H A D | mempolicy.c | 489 * Check if the folio's nid is in qp->nmask. 491 * If MPOL_MF_INVERT is set in qp->flags, check if the nid is 492 * in the invert of qp->nmask. 495 struct queue_pages *qp) 498 unsigned long flags = qp->flags; 500 return node_isset(nid, *qp->nmask) == !(flags & MPOL_MF_INVERT); 506 struct queue_pages *qp = walk->private; local 509 qp->nr_failed++; 517 if (!queue_folio_required(folio, qp)) 519 if (!(qp 494 queue_folio_required(struct folio *folio, struct queue_pages *qp) argument 540 struct queue_pages *qp = walk->private; local 620 struct queue_pages *qp = walk->private; local 692 struct queue_pages *qp = walk->private; local 763 struct queue_pages qp = { local [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 190 static int fill_ext_sge_inl_data(struct hns_roce_qp *qp, argument 194 struct ib_device *ibdev = &(to_hr_dev(qp->ibqp.device))->ib_dev; 202 if (msg_len > qp->sq.ext_sge_cnt * HNS_ROCE_SGE_SIZE) { 208 dseg = hns_roce_get_extend_sge(qp, idx & (qp->sge.sge_cnt - 1)); 238 dseg = hns_roce_get_extend_sge(qp, 239 idx & (qp->sge.sge_cnt - 1)); 249 static void set_extend_sge(struct hns_roce_qp *qp, struct ib_sge *sge, argument 256 dseg = hns_roce_get_extend_sge(qp, idx & (qp 268 check_inl_data_len(struct hns_roce_qp *qp, unsigned int len) argument 283 set_rc_inl(struct hns_roce_qp *qp, const struct ib_send_wr *wr, struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, unsigned int *sge_idx) argument 335 struct hns_roce_qp *qp = to_hr_qp(ibqp); local 462 set_ud_wqe(struct hns_roce_qp *qp, const struct ib_send_wr *wr, void *wqe, unsigned int *sge_idx, unsigned int owner_bit) argument 564 set_rc_wqe(struct hns_roce_qp *qp, const struct ib_send_wr *wr, void *wqe, unsigned int *sge_idx, unsigned int owner_bit) argument 615 update_sq_db(struct hns_roce_dev *hr_dev, struct hns_roce_qp *qp) argument 632 update_rq_db(struct hns_roce_dev *hr_dev, struct hns_roce_qp *qp) argument 668 write_dwqe(struct hns_roce_dev *hr_dev, struct hns_roce_qp *qp, void *wqe) argument 690 struct hns_roce_qp *qp = to_hr_qp(ibqp); local 2619 struct ib_qp *qp; local 2647 struct ib_qp *qp; local 3676 get_cqe_status(struct hns_roce_dev *hr_dev, struct hns_roce_qp *qp, struct hns_roce_cq *cq, struct hns_roce_v2_cqe *cqe, struct ib_wc *wc) argument 3890 struct hns_roce_qp *qp = *cur_qp; local [all...] |
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 278 void _pvrdma_flush_cqe(struct pvrdma_qp *qp, struct pvrdma_cq *cq) argument 306 if ((curr_cqe->qp & 0xFFFF) != qp->qp_handle) { 351 if (dev->qp_tbl[cqe->qp & 0xffff]) 352 *cur_qp = (struct pvrdma_qp *)dev->qp_tbl[cqe->qp & 0xffff]; 359 wc->qp = &(*cur_qp)->ibqp;
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager_cik.c | 262 struct queue_properties *qp, 278 m->cp_hqd_pq_doorbell_control = DOORBELL_OFFSET(qp->doorbell_off); 283 qp->is_active = 0; 300 struct queue_properties *qp, 313 qp->doorbell_off << SDMA0_RLC0_DOORBELL__OFFSET__SHIFT; 319 qp->is_active = 0; 260 restore_mqd(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument 298 restore_mqd_sdma(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument
|
/linux-master/drivers/crypto/hisilicon/hpre/ |
H A D | hpre_crypto.c | 118 struct hisi_qp *qp; member in struct:hpre_ctx 169 id = idr_alloc(&ctx->req_idr, NULL, 0, ctx->qp->sq_depth, GFP_ATOMIC); 219 struct hisi_qp *qp; local 222 qp = hpre_create_qp(type); 223 if (!qp) { 224 pr_err("Can not create hpre qp!\n"); 228 ret = hisi_qm_start_qp(qp, 0); 230 hisi_qm_free_qps(&qp, 1); 231 pci_err(qp->qm->pdev, "Can not start qp!\ 382 hpre_ctx_set(struct hpre_ctx *ctx, struct hisi_qp *qp, int qlen) argument 477 hpre_alg_cb(struct hisi_qp *qp, void *resp) argument 492 hpre_stop_qp_and_put(struct hisi_qp *qp) argument 500 struct hisi_qp *qp; local [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | type.h | 952 bool qp:1; member in struct:irdma_aeqe_info 1199 int irdma_sc_fast_register(struct irdma_sc_qp *qp, 1201 int irdma_sc_qp_create(struct irdma_sc_qp *qp, 1204 int irdma_sc_qp_destroy(struct irdma_sc_qp *qp, u64 scratch, 1206 int irdma_sc_qp_flush_wqes(struct irdma_sc_qp *qp, 1209 int irdma_sc_qp_init(struct irdma_sc_qp *qp, struct irdma_qp_init_info *info); 1210 int irdma_sc_qp_modify(struct irdma_sc_qp *qp, 1213 void irdma_sc_send_lsmm(struct irdma_sc_qp *qp, void *lsmm_buf, u32 size, 1216 void irdma_sc_send_rtt(struct irdma_sc_qp *qp, bool read); 1217 void irdma_sc_qp_setctx(struct irdma_sc_qp *qp, __le6 1232 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon934 1238 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon935 1244 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon936 1360 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon956 1366 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon957 1407 struct irdma_sc_qp *qp; member in struct:cqp_info::__anon933::__anon963 [all...] |
/linux-master/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 42 iser_err("qp event %s (%d)\n", 279 ib_conn->qp = ib_conn->cma_id->qp; 280 iser_info("setting conn %p cma_id %p qp %p max_send_wr %d\n", ib_conn, 281 ib_conn->cma_id, ib_conn->cma_id->qp, max_send_wr); 377 iser_info("freeing conn %p cma_id %p qp %p\n", 378 iser_conn, ib_conn->cma_id, ib_conn->qp); 380 if (ib_conn->qp) { 383 ib_conn->qp = NULL; 469 ib_drain_qp(ib_conn->qp); [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | ds.c | 997 struct ds_queue_entry *qp, *tmp; local 1005 list_for_each_entry_safe(qp, tmp, &todo, list) { 1006 struct ds_data *dpkt = (struct ds_data *) qp->req; 1007 struct ds_info *dp = qp->dp; 1009 int req_len = qp->req_len; 1023 list_del(&qp->list); 1024 kfree(qp); 1050 struct ds_queue_entry *qp; local 1052 qp = kmalloc(sizeof(struct ds_queue_entry) + len, GFP_ATOMIC); 1053 if (!qp) { [all...] |
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 224 event->event, ch, ch->sess_name, ch->qp->qp_num, 236 ch->sess_name, ch->qp->qp_num, 487 ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc, 848 return ib_post_recv(ch->qp, &wr, NULL); 872 ch->qp->qp_num); 874 return ib_post_send(ch->qp, &wr.wr, NULL); 879 struct srpt_rdma_ch *ch = wc->qp->qp_context; 881 pr_debug("%s-%d wc->status %d\n", ch->sess_name, ch->qp->qp_num, 891 ch->sess_name, ch->qp->qp_num); 925 ret = rdma_rw_ctx_init(&ctx->rw, ch->qp, c 1117 srpt_init_ch_qp(struct srpt_rdma_ch *ch, struct ib_qp *qp) argument 1157 srpt_ch_qp_rtr(struct srpt_rdma_ch *ch, struct ib_qp *qp) argument 1189 srpt_ch_qp_rts(struct srpt_rdma_ch *ch, struct ib_qp *qp) argument [all...] |
/linux-master/drivers/nvme/host/ |
H A D | rdma.c | 92 struct ib_qp *qp; member in struct:nvme_rdma_queue 281 queue->qp = queue->cm_id->qp; 433 ib_mr_pool_destroy(queue->qp, &queue->qp->sig_mrs); 434 ib_mr_pool_destroy(queue->qp, &queue->qp->rdma_mrs); 441 ib_destroy_qp(queue->qp); 528 ret = ib_mr_pool_init(queue->qp, &queue->qp [all...] |
/linux-master/drivers/media/pci/tw5864/ |
H A D | tw5864.h | 120 int qp; member in struct:tw5864_input 190 void tw5864_h264_put_stream_header(u8 **buf, size_t *space_left, int qp,
|