/linux-master/drivers/net/ethernet/amazon/ena/ |
H A D | ena_eth_com.h | 58 int ena_com_rx_pkt(struct ena_com_io_cq *io_cq, 66 bool ena_com_cq_empty(struct ena_com_io_cq *io_cq); 68 static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq, argument 71 writel(intr_reg->intr_control, io_cq->unmask_reg); 172 static inline void ena_com_update_numa_node(struct ena_com_io_cq *io_cq, argument 177 if (!io_cq->numa_node_cfg_reg) 183 writel(numa_cfg.numa_cfg, io_cq->numa_node_cfg_reg); 191 static inline void ena_com_cq_inc_head(struct ena_com_io_cq *io_cq) argument 193 io_cq->head++; 196 if (unlikely((io_cq 200 ena_com_tx_comp_req_id_get(struct ena_com_io_cq *io_cq, u16 *req_id) argument [all...] |
H A D | ena_eth_com.c | 9 struct ena_com_io_cq *io_cq) 15 head_masked = io_cq->head & (io_cq->q_depth - 1); 16 expected_phase = io_cq->phase; 18 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr 19 + (head_masked * io_cq->cdesc_entry_size_in_bytes)); 224 ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) argument 226 idx &= (io_cq->q_depth - 1); 228 ((uintptr_t)io_cq->cdesc_addr.virt_addr + 229 idx * io_cq 8 ena_com_get_next_rx_cdesc( struct ena_com_io_cq *io_cq) argument 232 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq, u16 *first_cdesc_idx) argument 342 ena_com_rx_set_flags(struct ena_com_io_cq *io_cq, struct ena_com_rx_ctx *ena_rx_ctx, struct ena_eth_io_rx_cdesc_base *cdesc) argument 532 ena_com_rx_pkt(struct ena_com_io_cq *io_cq, struct ena_com_io_sq *io_sq, struct ena_com_rx_ctx *ena_rx_ctx) argument 628 ena_com_cq_empty(struct ena_com_io_cq *io_cq) argument [all...] |
H A D | ena_com.c | 397 struct ena_com_io_cq *io_cq) 402 memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr)); 405 io_cq->cdesc_entry_size_in_bytes = 406 (io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ? 410 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; 414 io_cq->cdesc_addr.virt_addr = 415 dma_alloc_coherent(ena_dev->dmadev, size, &io_cq->cdesc_addr.phys_addr, GFP_KERNEL); 417 if (!io_cq 395 ena_com_init_io_cq(struct ena_com_dev *ena_dev, struct ena_com_create_io_ctx *ctx, struct ena_com_io_cq *io_cq) argument 917 ena_com_io_queue_free(struct ena_com_dev *ena_dev, struct ena_com_io_sq *io_sq, struct ena_com_io_cq *io_cq) argument 1346 ena_com_create_io_cq(struct ena_com_dev *ena_dev, struct ena_com_io_cq *io_cq) argument 1399 ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid, struct ena_com_io_sq **io_sq, struct ena_com_io_cq **io_cq) argument 1450 ena_com_destroy_io_cq(struct ena_com_dev *ena_dev, struct ena_com_io_cq *io_cq) argument 1813 struct ena_com_io_cq *io_cq; local 1872 struct ena_com_io_cq *io_cq; local [all...] |
H A D | ena_com.h | 443 * @io_cq - IO completion queue handler. 449 struct ena_com_io_cq **io_cq); 832 * @io_cq - io completion queue handler 839 struct ena_com_io_cq *io_cq); 843 * @io_cq - io completion queue handler 850 struct ena_com_io_cq *io_cq); 941 /* ena_com_io_cq_to_ena_dev - Extract ena_com_dev using contained field io_cq. 946 static inline struct ena_com_dev *ena_com_io_cq_to_ena_dev(struct ena_com_io_cq *io_cq) argument 948 return container_of(io_cq, struct ena_com_dev, io_cq_queues[io_cq [all...] |
/linux-master/include/linux/ |
H A D | iocontext.h | 15 * An io_cq (icq) is association between an io_context (ioc) and a 21 * io_cq and elevator can use the tail area for private information. The 22 * recommended way to do this is defining a struct which contains io_cq as 27 * struct io_cq icq; 73 struct io_cq { struct 78 * q_node and ioc_node link io_cq through icq_list of q and ioc 81 * used for RCU free of io_cq. 110 struct io_cq __rcu *icq_hint;
|
H A D | blk-mq.h | 175 struct io_cq *icq;
|
/linux-master/block/ |
H A D | blk-ioc.c | 37 struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); 46 static void ioc_exit_icq(struct io_cq *icq) 61 struct io_cq *icq; 73 static void ioc_destroy_icq(struct io_cq *icq) 119 struct io_cq *icq = hlist_entry(ioc->icq_list.first, 120 struct io_cq, ioc_node); 169 * Walk @q->icq_list and exit all io_cq's. 175 struct io_cq *icq = 176 list_first_entry(&q->icq_list, struct io_cq, q_nod [all...] |
H A D | elevator.h | 9 struct io_cq; 49 void (*init_icq)(struct io_cq *); 50 void (*exit_icq)(struct io_cq *);
|
H A D | elevator.c | 511 if (WARN_ON(e->icq_size < sizeof(struct io_cq)) || 512 WARN_ON(e->icq_align < __alignof__(struct io_cq)))
|
H A D | blk.h | 371 struct io_cq *ioc_find_get_icq(struct request_queue *q); 372 struct io_cq *ioc_lookup_icq(struct request_queue *q);
|
H A D | bfq-iosched.h | 478 /* associated io_cq structure */ 479 struct io_cq icq; /* must be the first member */
|
H A D | bfq-iosched.c | 445 static struct bfq_io_cq *icq_to_bic(struct io_cq *icq) 2794 * have the io_cq of this process. So we can immediately 2795 * configure this io_cq to redirect the requests of the 2796 * process to new_bfqq. In contrast, the io_cq of new_bfqq is 5466 static void bfq_exit_icq(struct io_cq *icq)
|
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 471 cq = phba->sli4_hba.hdwq[wqidx].io_cq; 648 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) 653 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq);
|
H A D | lpfc_sli4.h | 700 struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ member in struct:lpfc_sli4_hdw_queue
|
H A D | lpfc_debugfs.c | 3962 qp = phba->sli4_hba.hdwq[eqidx].io_cq; 4426 qp = phba->sli4_hba.hdwq[qidx].io_cq;
|
H A D | lpfc_init.c | 10404 phba->sli4_hba.hdwq[idx].io_cq = qdesc; 10851 lpfc_sli4_queue_free(hdwq[idx].io_cq); 10854 hdwq[idx].io_cq = NULL; 11191 qp[qidx].io_cq, 11500 lpfc_cq_destroy(phba, qp->io_cq);
|
H A D | lpfc_sli.c | 6162 sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0,
|