Searched refs:io_cq (Results 1 - 17 of 17) sorted by relevance

/linux-master/drivers/net/ethernet/amazon/ena/
H A Dena_eth_com.h58 int ena_com_rx_pkt(struct ena_com_io_cq *io_cq,
66 bool ena_com_cq_empty(struct ena_com_io_cq *io_cq);
68 static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq, argument
71 writel(intr_reg->intr_control, io_cq->unmask_reg);
172 static inline void ena_com_update_numa_node(struct ena_com_io_cq *io_cq, argument
177 if (!io_cq->numa_node_cfg_reg)
183 writel(numa_cfg.numa_cfg, io_cq->numa_node_cfg_reg);
191 static inline void ena_com_cq_inc_head(struct ena_com_io_cq *io_cq) argument
193 io_cq->head++;
196 if (unlikely((io_cq
200 ena_com_tx_comp_req_id_get(struct ena_com_io_cq *io_cq, u16 *req_id) argument
[all...]
H A Dena_eth_com.c9 struct ena_com_io_cq *io_cq)
15 head_masked = io_cq->head & (io_cq->q_depth - 1);
16 expected_phase = io_cq->phase;
18 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr
19 + (head_masked * io_cq->cdesc_entry_size_in_bytes));
224 ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) argument
226 idx &= (io_cq->q_depth - 1);
228 ((uintptr_t)io_cq->cdesc_addr.virt_addr +
229 idx * io_cq
8 ena_com_get_next_rx_cdesc( struct ena_com_io_cq *io_cq) argument
232 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq, u16 *first_cdesc_idx) argument
342 ena_com_rx_set_flags(struct ena_com_io_cq *io_cq, struct ena_com_rx_ctx *ena_rx_ctx, struct ena_eth_io_rx_cdesc_base *cdesc) argument
532 ena_com_rx_pkt(struct ena_com_io_cq *io_cq, struct ena_com_io_sq *io_sq, struct ena_com_rx_ctx *ena_rx_ctx) argument
628 ena_com_cq_empty(struct ena_com_io_cq *io_cq) argument
[all...]
H A Dena_com.c397 struct ena_com_io_cq *io_cq)
402 memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr));
405 io_cq->cdesc_entry_size_in_bytes =
406 (io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ?
410 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth;
414 io_cq->cdesc_addr.virt_addr =
415 dma_alloc_coherent(ena_dev->dmadev, size, &io_cq->cdesc_addr.phys_addr, GFP_KERNEL);
417 if (!io_cq
395 ena_com_init_io_cq(struct ena_com_dev *ena_dev, struct ena_com_create_io_ctx *ctx, struct ena_com_io_cq *io_cq) argument
917 ena_com_io_queue_free(struct ena_com_dev *ena_dev, struct ena_com_io_sq *io_sq, struct ena_com_io_cq *io_cq) argument
1346 ena_com_create_io_cq(struct ena_com_dev *ena_dev, struct ena_com_io_cq *io_cq) argument
1399 ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid, struct ena_com_io_sq **io_sq, struct ena_com_io_cq **io_cq) argument
1450 ena_com_destroy_io_cq(struct ena_com_dev *ena_dev, struct ena_com_io_cq *io_cq) argument
1813 struct ena_com_io_cq *io_cq; local
1872 struct ena_com_io_cq *io_cq; local
[all...]
H A Dena_com.h443 * @io_cq - IO completion queue handler.
449 struct ena_com_io_cq **io_cq);
832 * @io_cq - io completion queue handler
839 struct ena_com_io_cq *io_cq);
843 * @io_cq - io completion queue handler
850 struct ena_com_io_cq *io_cq);
941 /* ena_com_io_cq_to_ena_dev - Extract ena_com_dev using contained field io_cq.
946 static inline struct ena_com_dev *ena_com_io_cq_to_ena_dev(struct ena_com_io_cq *io_cq) argument
948 return container_of(io_cq, struct ena_com_dev, io_cq_queues[io_cq
[all...]
/linux-master/include/linux/
H A Diocontext.h15 * An io_cq (icq) is association between an io_context (ioc) and a
21 * io_cq and elevator can use the tail area for private information. The
22 * recommended way to do this is defining a struct which contains io_cq as
27 * struct io_cq icq;
73 struct io_cq { struct
78 * q_node and ioc_node link io_cq through icq_list of q and ioc
81 * used for RCU free of io_cq.
110 struct io_cq __rcu *icq_hint;
H A Dblk-mq.h175 struct io_cq *icq;
/linux-master/block/
H A Dblk-ioc.c37 struct io_cq *icq = container_of(head, struct io_cq, __rcu_head);
46 static void ioc_exit_icq(struct io_cq *icq)
61 struct io_cq *icq;
73 static void ioc_destroy_icq(struct io_cq *icq)
119 struct io_cq *icq = hlist_entry(ioc->icq_list.first,
120 struct io_cq, ioc_node);
169 * Walk @q->icq_list and exit all io_cq's.
175 struct io_cq *icq =
176 list_first_entry(&q->icq_list, struct io_cq, q_nod
[all...]
H A Delevator.h9 struct io_cq;
49 void (*init_icq)(struct io_cq *);
50 void (*exit_icq)(struct io_cq *);
H A Delevator.c511 if (WARN_ON(e->icq_size < sizeof(struct io_cq)) ||
512 WARN_ON(e->icq_align < __alignof__(struct io_cq)))
H A Dblk.h371 struct io_cq *ioc_find_get_icq(struct request_queue *q);
372 struct io_cq *ioc_lookup_icq(struct request_queue *q);
H A Dbfq-iosched.h478 /* associated io_cq structure */
479 struct io_cq icq; /* must be the first member */
H A Dbfq-iosched.c445 static struct bfq_io_cq *icq_to_bic(struct io_cq *icq)
2794 * have the io_cq of this process. So we can immediately
2795 * configure this io_cq to redirect the requests of the
2796 * process to new_bfqq. In contrast, the io_cq of new_bfqq is
5466 static void bfq_exit_icq(struct io_cq *icq)
/linux-master/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h471 cq = phba->sli4_hba.hdwq[wqidx].io_cq;
648 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid)
653 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq);
H A Dlpfc_sli4.h700 struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ member in struct:lpfc_sli4_hdw_queue
H A Dlpfc_debugfs.c3962 qp = phba->sli4_hba.hdwq[eqidx].io_cq;
4426 qp = phba->sli4_hba.hdwq[qidx].io_cq;
H A Dlpfc_init.c10404 phba->sli4_hba.hdwq[idx].io_cq = qdesc;
10851 lpfc_sli4_queue_free(hdwq[idx].io_cq);
10854 hdwq[idx].io_cq = NULL;
11191 qp[qidx].io_cq,
11500 lpfc_cq_destroy(phba, qp->io_cq);
H A Dlpfc_sli.c6162 sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0,

Completed in 311 milliseconds