Searched refs:scq (Results 1 - 23 of 23) sorted by relevance

/linux-master/drivers/atm/
H A Dnicstar.c27 * 2 - Per SCQ scq spinlock
115 #define scq_virt_to_bus(scq, p) \
116 (scq->dma + ((unsigned long)(p) - (unsigned long)(scq)->org))
126 static void free_scq(ns_dev *card, scq_info * scq, struct atm_vcc *vcc);
134 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd,
137 static void drain_scq(ns_dev * card, scq_info * scq, int pos);
248 free_scq(card, card->scd2vc[j]->scq, card->scd2vc[j]->tx_vcc);
864 scq_info *scq; local
869 scq
901 free_scq(ns_dev *card, scq_info *scq, struct atm_vcc *vcc) argument
1272 scq_info *scq; local
1466 scq_info *scq; local
1527 scq_info *scq = card->scq0; local
1626 scq_info *scq; local
1728 push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, struct sk_buff *skb, bool may_sleep) argument
1843 scq_info *scq; local
1915 drain_scq(ns_dev * card, scq_info * scq, int pos) argument
[all...]
H A Didt77252.c639 struct scq_info *scq; local
641 scq = kzalloc(sizeof(struct scq_info), GFP_KERNEL);
642 if (!scq)
644 scq->base = dma_alloc_coherent(&card->pcidev->dev, SCQ_SIZE,
645 &scq->paddr, GFP_KERNEL);
646 if (scq->base == NULL) {
647 kfree(scq);
651 scq->next = scq->base;
652 scq
668 free_scq(struct idt77252_dev *card, struct scq_info *scq) argument
705 struct scq_info *scq = vc->scq; local
793 struct scq_info *scq = vc->scq; local
950 fill_scd(struct idt77252_dev *card, struct scq_info *scq, int class) argument
959 clear_scd(struct idt77252_dev *card, struct scq_info *scq, int class) argument
[all...]
H A Dnicstar.h702 scq_info *scq; /* To keep track of the SCQ */ member in struct:vc_map
H A Didt77252.h217 struct scq_info *scq; /* To keep track of the SCQ */ member in struct:vc_map
/linux-master/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_qp.c65 static void pvrdma_lock_cqs(struct pvrdma_cq *scq, struct pvrdma_cq *rcq,
68 __acquires(scq->cq_lock) __acquires(rcq->cq_lock)
70 if (scq == rcq) {
71 spin_lock_irqsave(&scq->cq_lock, *scq_flags);
73 } else if (scq->cq_handle < rcq->cq_handle) {
74 spin_lock_irqsave(&scq->cq_lock, *scq_flags);
79 spin_lock_irqsave_nested(&scq->cq_lock, *scq_flags,
84 static void pvrdma_unlock_cqs(struct pvrdma_cq *scq, struct pvrdma_cq *rcq,
87 __releases(scq->cq_lock) __releases(rcq->cq_lock)
89 if (scq
103 struct pvrdma_cq *scq, *rcq; local
453 struct pvrdma_cq *scq; local
[all...]
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_qp.c357 struct rxe_cq *scq = to_rcq(init->send_cq); local
363 rxe_get(scq);
369 qp->scq = scq;
373 atomic_inc(&scq->num_wq);
397 atomic_dec(&scq->num_wq);
401 qp->scq = NULL;
406 rxe_put(scq);
847 if (qp->scq) {
848 atomic_dec(&qp->scq
[all...]
H A Drxe_comp.c466 rxe_cq_post(qp->scq, &cqe, 0);
583 err = rxe_cq_post(qp->scq, &cqe, 0);
585 rxe_dbg_cq(qp->scq, "post cq failed, err = %d\n", err);
H A Drxe_verbs.h212 struct rxe_cq *scq; member in struct:rxe_qp
/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c70 struct bnxt_qplib_cq *scq, *rcq; local
72 scq = qp->scq;
76 dev_dbg(&scq->hwq.pdev->dev,
79 list_add_tail(&qp->sq_flush, &scq->sqf_head);
94 __acquires(&qp->scq->flush_lock) __acquires(&qp->rcq->flush_lock)
96 spin_lock_irqsave(&qp->scq->flush_lock, *flags);
97 if (qp->scq == qp->rcq)
105 __releases(&qp->scq->flush_lock) __releases(&qp->rcq->flush_lock)
107 if (qp->scq
[all...]
H A Dib_verbs.h94 struct bnxt_re_cq *scq; member in struct:bnxt_re_qp
H A Dib_verbs.c820 __acquires(&qp->scq->cq_lock) __acquires(&qp->rcq->cq_lock)
824 spin_lock_irqsave(&qp->scq->cq_lock, flags);
825 if (qp->rcq != qp->scq)
835 __releases(&qp->scq->cq_lock) __releases(&qp->rcq->cq_lock)
837 if (qp->rcq != qp->scq)
841 spin_unlock_irqrestore(&qp->scq->cq_lock, flags);
936 scq_nq = qplib_qp->scq->nq;
1165 qp->qplib_qp.scq = qp1_qp->scq;
1396 qplqp->scq
[all...]
H A Dqplib_fp.h311 struct bnxt_qplib_cq *scq; member in struct:bnxt_qplib_qp
437 * Each flush list is protected by qplib_cq->flush_lock. Both scq and rcq
/linux-master/drivers/infiniband/hw/cxgb4/
H A Dev.c139 cqid = qhp->attr.scq;
H A Dqp.c200 struct t4_cq *rcq, struct t4_cq *scq,
338 FW_RI_RES_WR_IQID_V(scq->cqid));
1812 wqe->u.init.scqid = cpu_to_be32(qhp->attr.scq);
2188 qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid;
199 create_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, struct t4_cq *rcq, struct t4_cq *scq, struct c4iw_dev_ucontext *uctx, struct c4iw_wr_wait *wr_waitp, int need_rq) argument
H A Diw_cxgb4.h450 u32 scq; member in struct:c4iw_qp_attributes
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_verbs.h233 struct erdma_cq *scq; member in struct:erdma_qp
H A Derdma_verbs.c64 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->scq->cqn);
89 FIELD_PREP(ERDMA_CMD_CREATE_QP_CQN_MASK, qp->scq->cqn);
938 qp->scq = to_ecq(attrs->send_cq);
/linux-master/drivers/infiniband/core/
H A Duverbs_cmd.c1286 struct ib_cq *scq = NULL, *rcq = NULL; local
1383 scq = uobj_get_obj_read(cq, UVERBS_OBJECT_CQ,
1386 rcq = rcq ?: scq;
1389 if (!pd || (!scq && has_sq)) {
1398 attr.send_cq = scq;
1459 if (scq)
1460 rdma_lookup_put_uobject(&scq->uobject->uevent.uobject,
1462 if (rcq && rcq != scq)
1487 if (scq)
1488 rdma_lookup_put_uobject(&scq
[all...]
/linux-master/drivers/infiniband/hw/hns/
H A Dhns_roce_qp.c182 struct ib_cq *scq = init_attr->send_cq; local
189 if (scq)
190 cqn = to_hr_cq(scq)->cqn;
193 if (scq && (get_affinity_cq_bank(i) != (cqn & CQ_BANKID_MASK)))
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Dresource_tracker.c112 struct res_cq *scq; member in struct:res_qp
2966 struct res_cq *scq; local
3004 err = get_res(dev, slave, scqn, RES_CQ, &scq);
3008 scq = rcq;
3025 atomic_inc(&scq->ref_count);
3026 qp->scq = scq;
4004 atomic_dec(&qp->scq->ref_count);
4717 atomic_dec(&qp->scq->ref_count);
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw.h430 struct siw_cq *scq; member in struct:siw_qp
H A Dsiw_verbs.c397 qp->scq = to_siw_cq(attrs->send_cq);
625 qp->scq = qp->rcq = NULL;
H A Dsiw_qp.c1066 struct siw_cq *cq = qp->scq;

Completed in 597 milliseconds