Searched refs:cq (Results 1 - 25 of 350) sorted by path

1234567891011>>

/linux-master/drivers/infiniband/hw/cxgb4/
H A DMakefile7 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o \
/linux-master/drivers/infiniband/hw/mlx4/
H A DMakefile4 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
/linux-master/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.h125 void ocrdma_mbx_destroy_cq(struct ocrdma_dev *dev, struct ocrdma_cq *cq);
/linux-master/drivers/infiniband/sw/rdmavt/
H A DMakefile11 rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
/linux-master/drivers/isdn/hardware/mISDN/
H A Dspeedfax.c208 channel_ctrl(struct sfax_hw *sf, struct mISDN_ctrl_req *cq) argument
212 switch (cq->op) {
214 cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3;
217 /* cq->channel: 0 disable, 1 B1 loop 2 B2 loop, 3 both */
218 if (cq->channel < 0 || cq->channel > 3) {
222 ret = sf->isac.ctrl(&sf->isac, HW_TESTLOOP, cq->channel);
225 ret = sf->isac.ctrl(&sf->isac, HW_TIMER3_VALUE, cq->p1);
228 pr_info("%s: unknown Op %x\n", sf->name, cq->op);
/linux-master/drivers/isdn/mISDN/
H A Ddsp_hwec.c38 struct mISDN_ctrl_req cq; local
82 memset(&cq, 0, sizeof(cq));
83 cq.op = MISDN_CTRL_HFC_ECHOCAN_ON;
84 cq.p1 = deftaps;
85 if (!dsp->ch.peer->ctrl(&dsp->ch, CONTROL_CHANNEL, &cq)) {
94 struct mISDN_ctrl_req cq; local
103 memset(&cq, 0, sizeof(cq));
104 cq
[all...]
/linux-master/drivers/net/ethernet/brocade/bna/
H A Dbfi_enet.h495 struct bfi_enet_cq cq; member in struct:bfi_enet_rx_cfg_req::__anon473
/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_hw.h232 static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes) argument
234 struct h_epa epa = cq->epas.kernel;
239 static inline void ehea_reset_cq_n1(struct ehea_cq *cq) argument
241 struct h_epa epa = cq->epas.kernel;
H A Dehea_qmr.h333 static inline void ehea_inc_cq(struct ehea_cq *cq) argument
335 hw_qeit_inc(&cq->hw_queue);
367 int ehea_destroy_cq(struct ehea_cq *cq);
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A DMakefile4 mlx4_core-y := alloc.o catas.o cmd.o cq.o eq.o fw.o fw_qos.o icm.o intf.o \
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_dim.c51 mlx5e_complete_dim_work(dim, cur_moder, rq->mdev, &rq->cq.mcq);
61 mlx5e_complete_dim_work(dim, cur_moder, sq->cq.mdev, &sq->cq.mcq);
/linux-master/drivers/scsi/be2iscsi/
H A Dbe.h91 struct be_queue_info *cq; member in struct:be_eq_obj
98 struct be_queue_info cq; member in struct:be_mcc_obj
/linux-master/drivers/accel/habanalabs/common/
H A Dhw_queue.c115 * @reserve_cq_entry : whether to reserve an entry in the cq
284 struct hl_cq *cq; local
318 cq = &hdev->completion_queue[q->cq_id];
319 cq_addr = cq->bus_address + cq->pi * sizeof(struct hl_cq_entry);
330 cq->pi = hl_cq_inc_ptr(cq->pi);
H A Dirq.c27 * hl_cq_inc_ptr - increment ci or pi of cq
74 * @cq: completion queue
78 static void job_finish(struct hl_device *hdev, u32 cs_seq, struct hl_cq *cq, ktime_t timestamp) argument
83 queue = &hdev->kernel_queues[cq->hw_queue_id];
86 queue_work(hdev->cq_wq[cq->cq_idx], &job->finish_work);
132 struct hl_cq *cq = arg; local
133 struct hl_device *hdev = cq->hdev;
142 irq, cq->hw_queue_id);
146 cq_base = cq->kernel_address;
149 cq_entry = (struct hl_cq_entry *) &cq_base[cq
[all...]
/linux-master/drivers/accel/habanalabs/gaudi2/
H A Dgaudi2.c4311 struct hl_cq *cq; local
4327 cq = &hdev->completion_queue[GAUDI2_RESERVED_CQ_CS_COMPLETION];
4328 rc = request_irq(irq, hl_irq_handler_cq, 0, gaudi2_irq_name(GAUDI2_IRQ_NUM_COMPLETION), cq);
4410 free_irq(irq, cq);
4414 free_irq(irq, cq);
4456 struct hl_cq *cq; local
4484 cq = &hdev->completion_queue[GAUDI2_RESERVED_CQ_CS_COMPLETION];
4485 free_irq(irq, cq);
6948 struct hl_cq *cq; local
6980 cq
[all...]
/linux-master/drivers/dma/
H A Dfsl-qdma.c199 struct fsl_qdma_format *cq; member in struct:fsl_qdma_queue
517 queue_temp->cq =
523 if (!queue_temp->cq)
529 queue_temp->virt_head = queue_temp->cq;
530 queue_temp->virt_tail = queue_temp->cq;
567 status_head->cq = dmam_alloc_coherent(&pdev->dev,
572 if (!status_head->cq)
576 status_head->virt_head = status_head->cq;
577 status_head->virt_tail = status_head->cq;
684 if (fsl_status->virt_head == fsl_status->cq
[all...]
H A Dhisi_dma.c148 struct hisi_dma_cqe *cq; member in struct:hisi_dma_chan
470 memset(chan->cq, 0, sizeof(struct hisi_dma_cqe) * hdma_dev->chan_depth);
598 chan->cq = dmam_alloc_coherent(dev, cq_size, &chan->cq_dma,
600 if (!chan->cq)
615 /* set sq, cq base */
625 /* set sq, cq depth */
629 /* init sq tail and cq head */
738 cqe = chan->cq + chan->cq_head;
/linux-master/drivers/infiniband/core/
H A DMakefile10 ib_core-y := packer.o ud_header.o verbs.o cq.o rw.o sysfs.o \
H A Dcq.c42 struct ib_cq *cq = dim->priv; local
49 trace_cq_modify(cq, comps, usec);
50 cq->device->ops.modify_cq(cq, comps, usec);
53 static void rdma_dim_init(struct ib_cq *cq) argument
57 if (!cq->device->ops.modify_cq || !cq->device->use_cq_dim ||
58 cq->poll_ctx == IB_POLL_DIRECT)
68 dim->priv = cq;
69 cq
74 rdma_dim_destroy(struct ib_cq *cq) argument
83 __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) argument
92 __ib_process_cq(struct ib_cq *cq, int budget, struct ib_wc *wcs, int batch) argument
138 ib_process_cq_direct(struct ib_cq *cq, int budget) argument
146 ib_cq_completion_direct(struct ib_cq *cq, void *private) argument
153 struct ib_cq *cq = container_of(iop, struct ib_cq, iop); local
172 ib_cq_completion_softirq(struct ib_cq *cq, void *private) argument
180 struct ib_cq *cq = container_of(work, struct ib_cq, work); local
192 ib_cq_completion_workqueue(struct ib_cq *cq, void *private) argument
220 struct ib_cq *cq; local
318 ib_free_cq(struct ib_cq *cq) argument
353 struct ib_cq *cq, *n; local
372 struct ib_cq *cq, *n; local
434 struct ib_cq *cq, *found = NULL; local
498 ib_cq_pool_put(struct ib_cq *cq, unsigned int nr_cqe) argument
[all...]
H A Dmad.c109 static void ib_mad_send_done(struct ib_cq *cq, struct ib_wc *wc);
2017 static void ib_mad_recv_done(struct ib_cq *cq, struct ib_wc *wc) argument
2019 struct ib_mad_port_private *port_priv = cq->cq_context;
2266 static void ib_mad_send_done(struct ib_cq *cq, struct ib_wc *wc) argument
2268 struct ib_mad_port_private *port_priv = cq->cq_context;
2835 ret = ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP);
2894 qp_init_attr.send_cq = qp_info->port_priv->cq;
2895 qp_init_attr.recv_cq = qp_info->port_priv->cq;
2973 port_priv->cq = ib_alloc_cq(port_priv->device, port_priv, cq_size, 0,
2975 if (IS_ERR(port_priv->cq)) {
[all...]
H A Dmad_priv.h202 struct ib_cq *cq; member in struct:ib_mad_port_private
H A Dnldev.c406 [RDMA_RESTRACK_CQ] = "cq",
614 struct ib_cq *cq = container_of(res, struct ib_cq, res); local
615 struct ib_device *dev = cq->device;
617 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_CQE, cq->cqe))
620 atomic_read(&cq->usecnt), RDMA_NLDEV_ATTR_PAD))
625 nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_POLL_CTX, cq->poll_ctx))
628 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_DEV_DIM, (cq->dim != NULL)))
635 cq->uobject->uevent.uobject.context->res.id))
642 dev->ops.fill_res_cq_entry(msg, cq) : 0;
648 struct ib_cq *cq local
[all...]
H A Duverbs.h236 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context);
H A Duverbs_cmd.c1009 struct ib_cq *cq; local
1039 cq = rdma_zalloc_drv_obj(ib_dev, ib_cq);
1040 if (!cq) {
1044 cq->device = ib_dev;
1045 cq->uobject = obj;
1046 cq->comp_handler = ib_uverbs_comp_handler;
1047 cq->event_handler = ib_uverbs_cq_event_handler;
1048 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL;
1049 atomic_set(&cq->usecnt, 0);
1051 rdma_restrack_new(&cq
1122 struct ib_cq *cq; local
1183 struct ib_cq *cq; local
1233 struct ib_cq *cq; local
2887 struct ib_cq *cq; local
3656 struct ib_cq *cq; local
[all...]
H A Duverbs_main.c368 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context) argument
390 uobj = cq->uobject;
392 entry->desc.comp.cq_handle = cq->uobject->uevent.uobject.user_handle;
449 uverbs_uobj_event(&event->element.cq->uobject->uevent, event);

Completed in 489 milliseconds

1234567891011>>