/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | Makefile | 7 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o \
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | Makefile | 4 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
|
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.h | 125 void ocrdma_mbx_destroy_cq(struct ocrdma_dev *dev, struct ocrdma_cq *cq);
|
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | Makefile | 11 rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
|
/linux-master/drivers/isdn/hardware/mISDN/ |
H A D | speedfax.c | 208 channel_ctrl(struct sfax_hw *sf, struct mISDN_ctrl_req *cq) argument 212 switch (cq->op) { 214 cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3; 217 /* cq->channel: 0 disable, 1 B1 loop 2 B2 loop, 3 both */ 218 if (cq->channel < 0 || cq->channel > 3) { 222 ret = sf->isac.ctrl(&sf->isac, HW_TESTLOOP, cq->channel); 225 ret = sf->isac.ctrl(&sf->isac, HW_TIMER3_VALUE, cq->p1); 228 pr_info("%s: unknown Op %x\n", sf->name, cq->op);
|
/linux-master/drivers/isdn/mISDN/ |
H A D | dsp_hwec.c | 38 struct mISDN_ctrl_req cq; local 82 memset(&cq, 0, sizeof(cq)); 83 cq.op = MISDN_CTRL_HFC_ECHOCAN_ON; 84 cq.p1 = deftaps; 85 if (!dsp->ch.peer->ctrl(&dsp->ch, CONTROL_CHANNEL, &cq)) { 94 struct mISDN_ctrl_req cq; local 103 memset(&cq, 0, sizeof(cq)); 104 cq [all...] |
/linux-master/drivers/net/ethernet/brocade/bna/ |
H A D | bfi_enet.h | 495 struct bfi_enet_cq cq; member in struct:bfi_enet_rx_cfg_req::__anon473
|
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_hw.h | 232 static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes) argument 234 struct h_epa epa = cq->epas.kernel; 239 static inline void ehea_reset_cq_n1(struct ehea_cq *cq) argument 241 struct h_epa epa = cq->epas.kernel;
|
H A D | ehea_qmr.h | 333 static inline void ehea_inc_cq(struct ehea_cq *cq) argument 335 hw_qeit_inc(&cq->hw_queue); 367 int ehea_destroy_cq(struct ehea_cq *cq);
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | Makefile | 4 mlx4_core-y := alloc.o catas.o cmd.o cq.o eq.o fw.o fw_qos.o icm.o intf.o \
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_dim.c | 51 mlx5e_complete_dim_work(dim, cur_moder, rq->mdev, &rq->cq.mcq); 61 mlx5e_complete_dim_work(dim, cur_moder, sq->cq.mdev, &sq->cq.mcq);
|
/linux-master/drivers/scsi/be2iscsi/ |
H A D | be.h | 91 struct be_queue_info *cq; member in struct:be_eq_obj 98 struct be_queue_info cq; member in struct:be_mcc_obj
|
/linux-master/drivers/accel/habanalabs/common/ |
H A D | hw_queue.c | 115 * @reserve_cq_entry : whether to reserve an entry in the cq 284 struct hl_cq *cq; local 318 cq = &hdev->completion_queue[q->cq_id]; 319 cq_addr = cq->bus_address + cq->pi * sizeof(struct hl_cq_entry); 330 cq->pi = hl_cq_inc_ptr(cq->pi);
|
H A D | irq.c | 27 * hl_cq_inc_ptr - increment ci or pi of cq 74 * @cq: completion queue 78 static void job_finish(struct hl_device *hdev, u32 cs_seq, struct hl_cq *cq, ktime_t timestamp) argument 83 queue = &hdev->kernel_queues[cq->hw_queue_id]; 86 queue_work(hdev->cq_wq[cq->cq_idx], &job->finish_work); 132 struct hl_cq *cq = arg; local 133 struct hl_device *hdev = cq->hdev; 142 irq, cq->hw_queue_id); 146 cq_base = cq->kernel_address; 149 cq_entry = (struct hl_cq_entry *) &cq_base[cq [all...] |
/linux-master/drivers/accel/habanalabs/gaudi2/ |
H A D | gaudi2.c | 4311 struct hl_cq *cq; local 4327 cq = &hdev->completion_queue[GAUDI2_RESERVED_CQ_CS_COMPLETION]; 4328 rc = request_irq(irq, hl_irq_handler_cq, 0, gaudi2_irq_name(GAUDI2_IRQ_NUM_COMPLETION), cq); 4410 free_irq(irq, cq); 4414 free_irq(irq, cq); 4456 struct hl_cq *cq; local 4484 cq = &hdev->completion_queue[GAUDI2_RESERVED_CQ_CS_COMPLETION]; 4485 free_irq(irq, cq); 6948 struct hl_cq *cq; local 6980 cq [all...] |
/linux-master/drivers/dma/ |
H A D | fsl-qdma.c | 199 struct fsl_qdma_format *cq; member in struct:fsl_qdma_queue 517 queue_temp->cq = 523 if (!queue_temp->cq) 529 queue_temp->virt_head = queue_temp->cq; 530 queue_temp->virt_tail = queue_temp->cq; 567 status_head->cq = dmam_alloc_coherent(&pdev->dev, 572 if (!status_head->cq) 576 status_head->virt_head = status_head->cq; 577 status_head->virt_tail = status_head->cq; 684 if (fsl_status->virt_head == fsl_status->cq [all...] |
H A D | hisi_dma.c | 148 struct hisi_dma_cqe *cq; member in struct:hisi_dma_chan 470 memset(chan->cq, 0, sizeof(struct hisi_dma_cqe) * hdma_dev->chan_depth); 598 chan->cq = dmam_alloc_coherent(dev, cq_size, &chan->cq_dma, 600 if (!chan->cq) 615 /* set sq, cq base */ 625 /* set sq, cq depth */ 629 /* init sq tail and cq head */ 738 cqe = chan->cq + chan->cq_head;
|
/linux-master/drivers/infiniband/core/ |
H A D | Makefile | 10 ib_core-y := packer.o ud_header.o verbs.o cq.o rw.o sysfs.o \
|
H A D | cq.c | 42 struct ib_cq *cq = dim->priv; local 49 trace_cq_modify(cq, comps, usec); 50 cq->device->ops.modify_cq(cq, comps, usec); 53 static void rdma_dim_init(struct ib_cq *cq) argument 57 if (!cq->device->ops.modify_cq || !cq->device->use_cq_dim || 58 cq->poll_ctx == IB_POLL_DIRECT) 68 dim->priv = cq; 69 cq 74 rdma_dim_destroy(struct ib_cq *cq) argument 83 __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) argument 92 __ib_process_cq(struct ib_cq *cq, int budget, struct ib_wc *wcs, int batch) argument 138 ib_process_cq_direct(struct ib_cq *cq, int budget) argument 146 ib_cq_completion_direct(struct ib_cq *cq, void *private) argument 153 struct ib_cq *cq = container_of(iop, struct ib_cq, iop); local 172 ib_cq_completion_softirq(struct ib_cq *cq, void *private) argument 180 struct ib_cq *cq = container_of(work, struct ib_cq, work); local 192 ib_cq_completion_workqueue(struct ib_cq *cq, void *private) argument 220 struct ib_cq *cq; local 318 ib_free_cq(struct ib_cq *cq) argument 353 struct ib_cq *cq, *n; local 372 struct ib_cq *cq, *n; local 434 struct ib_cq *cq, *found = NULL; local 498 ib_cq_pool_put(struct ib_cq *cq, unsigned int nr_cqe) argument [all...] |
H A D | mad.c | 109 static void ib_mad_send_done(struct ib_cq *cq, struct ib_wc *wc); 2017 static void ib_mad_recv_done(struct ib_cq *cq, struct ib_wc *wc) argument 2019 struct ib_mad_port_private *port_priv = cq->cq_context; 2266 static void ib_mad_send_done(struct ib_cq *cq, struct ib_wc *wc) argument 2268 struct ib_mad_port_private *port_priv = cq->cq_context; 2835 ret = ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP); 2894 qp_init_attr.send_cq = qp_info->port_priv->cq; 2895 qp_init_attr.recv_cq = qp_info->port_priv->cq; 2973 port_priv->cq = ib_alloc_cq(port_priv->device, port_priv, cq_size, 0, 2975 if (IS_ERR(port_priv->cq)) { [all...] |
H A D | mad_priv.h | 202 struct ib_cq *cq; member in struct:ib_mad_port_private
|
H A D | nldev.c | 406 [RDMA_RESTRACK_CQ] = "cq", 614 struct ib_cq *cq = container_of(res, struct ib_cq, res); local 615 struct ib_device *dev = cq->device; 617 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_CQE, cq->cqe)) 620 atomic_read(&cq->usecnt), RDMA_NLDEV_ATTR_PAD)) 625 nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_POLL_CTX, cq->poll_ctx)) 628 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_DEV_DIM, (cq->dim != NULL))) 635 cq->uobject->uevent.uobject.context->res.id)) 642 dev->ops.fill_res_cq_entry(msg, cq) : 0; 648 struct ib_cq *cq local [all...] |
H A D | uverbs.h | 236 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context);
|
H A D | uverbs_cmd.c | 1009 struct ib_cq *cq; local 1039 cq = rdma_zalloc_drv_obj(ib_dev, ib_cq); 1040 if (!cq) { 1044 cq->device = ib_dev; 1045 cq->uobject = obj; 1046 cq->comp_handler = ib_uverbs_comp_handler; 1047 cq->event_handler = ib_uverbs_cq_event_handler; 1048 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; 1049 atomic_set(&cq->usecnt, 0); 1051 rdma_restrack_new(&cq 1122 struct ib_cq *cq; local 1183 struct ib_cq *cq; local 1233 struct ib_cq *cq; local 2887 struct ib_cq *cq; local 3656 struct ib_cq *cq; local [all...] |
H A D | uverbs_main.c | 368 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context) argument 390 uobj = cq->uobject; 392 entry->desc.comp.cq_handle = cq->uobject->uevent.uobject.user_handle; 449 uverbs_uobj_event(&event->element.cq->uobject->uevent, event);
|