/linux-master/drivers/net/wireless/broadcom/b43/ |
H A D | dma.h | 199 struct b43_dmadesc_generic *(*idx2desc) (struct b43_dmaring * ring, 203 void (*fill_descriptor) (struct b43_dmaring * ring, 207 void (*poke_tx) (struct b43_dmaring * ring, int slot); 208 void (*tx_suspend) (struct b43_dmaring * ring); 209 void (*tx_resume) (struct b43_dmaring * ring); 210 int (*get_current_rxslot) (struct b43_dmaring * ring); 211 void (*set_current_rxslot) (struct b43_dmaring * ring, int slot); 229 /* Kernel virtual base address of the ring memory. */ 235 * This is NULL for an RX ring. 238 /* (Unadjusted) DMA base bus-address of the ring memor 278 b43_dma_read(struct b43_dmaring *ring, u16 offset) argument 283 b43_dma_write(struct b43_dmaring *ring, u16 offset, u32 value) argument [all...] |
/linux-master/tools/testing/vsock/ |
H A D | vsock_uring_test.c | 64 struct io_uring ring; local 80 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring, 0)) 83 if (io_uring_register_buffers(&ring, iovec, test_data->vecs_cnt)) 89 sqe = io_uring_get_sqe(&ring); 96 if (io_uring_submit(&ring) != 1) 99 if (io_uring_wait_cqe(&ring, &cqe)) 102 io_uring_cqe_seen(&ring, cqe); 107 io_uring_queue_exit(&ring); 117 struct io_uring ring; local 137 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring, [all...] |
/linux-master/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_main.c | 390 static __le64 *xgene_enet_get_exp_bufs(struct xgene_enet_desc_ring *ring) argument 394 exp_bufs = &ring->exp_bufs[ring->exp_buf_tail * MAX_EXP_BUFFS]; 396 ring->exp_buf_tail = (ring->exp_buf_tail + 1) & ((ring->slots / 2) - 1); 401 static dma_addr_t *xgene_get_frag_dma_array(struct xgene_enet_desc_ring *ring) argument 403 return &ring->cp_ring->frag_dma_addr[ring->tail * MAX_SKB_FRAGS]; 783 static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring, argument 850 struct xgene_enet_desc_ring *ring; local 882 struct xgene_enet_desc_ring *ring; local 907 struct xgene_enet_desc_ring *ring; local 939 struct xgene_enet_desc_ring *ring; local 1048 xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring) argument 1063 struct xgene_enet_desc_ring *ring; local 1129 xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring) argument 1151 struct xgene_enet_desc_ring *ring; local 1195 is_irq_mbox_required(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument 1206 xgene_enet_ring_cmd_base(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument 1220 struct xgene_enet_desc_ring *ring; local 1477 struct xgene_enet_desc_ring *ring; local [all...] |
/linux-master/drivers/scsi/snic/ |
H A D | vnic_dev.c | 176 unsigned int svnic_dev_desc_ring_size(struct vnic_dev_ring *ring, argument 189 ring->base_align = 512; 194 ring->desc_count = ALIGN(desc_count, count_align); 196 ring->desc_size = ALIGN(desc_size, desc_align); 198 ring->size = ring->desc_count * ring->desc_size; 199 ring->size_unaligned = ring->size + ring 204 svnic_dev_clear_desc_ring(struct vnic_dev_ring *ring) argument 209 svnic_dev_alloc_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring, unsigned int desc_count, unsigned int desc_size) argument 236 svnic_dev_free_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring) argument [all...] |
H A D | vnic_cq.h | 41 struct vnic_dev_ring ring; member in struct:vnic_cq 57 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 58 cq->ring.desc_size * cq->to_clean); 69 if (cq->to_clean == cq->ring.desc_count) { 74 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 75 cq->ring.desc_size * cq->to_clean);
|
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | adreno_gpu.c | 569 struct msm_ringbuffer *ring = gpu->rb[i]; local 571 if (!ring) 574 ring->cur = ring->start; 575 ring->next = ring->start; 576 ring->memptrs->rptr = 0; 577 ring->memptrs->bv_fence = ring->fctx->completed_fence; 583 if (fence_before(ring 592 get_rptr(struct adreno_gpu *adreno_gpu, struct msm_ringbuffer *ring) argument 623 adreno_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring, u32 reg) argument 643 adreno_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument 925 struct msm_ringbuffer *ring = gpu->rb[i]; local 959 ring_freewords(struct msm_ringbuffer *ring) argument 969 adreno_wait_ring(struct msm_ringbuffer *ring, uint32_t ndwords) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | vcn_v3_0.c | 77 static void vcn_v3_0_dec_ring_set_wptr(struct amdgpu_ring *ring); 78 static void vcn_v3_0_enc_ring_set_wptr(struct amdgpu_ring *ring); 85 * Set ring and irq function pointers 126 struct amdgpu_ring *ring; local 186 ring = &adev->vcn.inst[i].ring_dec; 187 ring->use_doorbell = true; 189 ring->doorbell_index = vcn_doorbell_index + i * (adev->vcn.num_enc_rings + 1); 191 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 8 * i; 193 ring->vm_hub = AMDGPU_MMHUB0(0); 194 sprintf(ring 300 struct amdgpu_ring *ring; local 948 struct amdgpu_ring *ring; local 1100 struct amdgpu_ring *ring; local 1289 struct amdgpu_ring *ring; local 1605 struct amdgpu_ring *ring; local 1689 vcn_v3_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 1703 vcn_v3_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 1720 vcn_v3_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument 1874 struct amdgpu_ring *ring = amdgpu_job_ring(job); local 1940 vcn_v3_0_enc_ring_get_rptr(struct amdgpu_ring *ring) argument 1957 vcn_v3_0_enc_ring_get_wptr(struct amdgpu_ring *ring) argument 1981 vcn_v3_0_enc_ring_set_wptr(struct amdgpu_ring *ring) argument [all...] |
H A D | psp_v11_0_8.c | 38 /* Write the ring destroy command*/ 47 /* Write the ring destroy command*/ 65 struct psp_ring *ring = &psp->km_ring; local 75 /* Write low address of the ring to C2PMSG_102 */ 76 psp_ring_reg = lower_32_bits(ring->ring_mem_mc_addr); 78 /* Write high address of the ring to C2PMSG_103 */ 79 psp_ring_reg = upper_32_bits(ring->ring_mem_mc_addr); 82 /* Write the ring initialization command to C2PMSG_101 */ 94 /* Wait for sOS ready for ring creation */ 98 DRM_ERROR("Failed to wait for trust OS ready for ring creatio 131 struct psp_ring *ring = &psp->km_ring; local [all...] |
H A D | gfx_v8_0.c | 730 static void gfx_v8_0_ring_emit_ce_meta(struct amdgpu_ring *ring); 731 static void gfx_v8_0_ring_emit_de_meta(struct amdgpu_ring *ring); 838 static int gfx_v8_0_ring_test_ring(struct amdgpu_ring *ring) argument 840 struct amdgpu_device *adev = ring->adev; 846 r = amdgpu_ring_alloc(ring, 3); 850 amdgpu_ring_write(ring, PACKET3(PACKET3_SET_UCONFIG_REG, 1)); 851 amdgpu_ring_write(ring, mmSCRATCH_REG0 - PACKET3_SET_UCONFIG_REG_START); 852 amdgpu_ring_write(ring, 0xDEADBEEF); 853 amdgpu_ring_commit(ring); 868 static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, lon argument 1490 struct amdgpu_ring *ring = &adev->gfx.compute_ring[0]; local 1865 struct amdgpu_ring *ring = &adev->gfx.compute_ring[ring_id]; local 1904 struct amdgpu_ring *ring; local 4141 struct amdgpu_ring *ring = &adev->gfx.gfx_ring[0]; local 4202 gfx_v8_0_set_cpg_door_bell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 4238 struct amdgpu_ring *ring; local 4301 gfx_v8_0_kiq_setting(struct amdgpu_ring *ring) argument 4351 struct amdgpu_ring *ring = &adev->gfx.compute_ring[i]; local 4397 gfx_v8_0_mqd_set_priority(struct amdgpu_ring *ring, struct vi_mqd *mqd) argument 4410 gfx_v8_0_mqd_init(struct amdgpu_ring *ring) argument 4594 gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring) argument 4634 gfx_v8_0_kcq_init_queue(struct amdgpu_ring *ring) argument 4675 struct amdgpu_ring *ring; local 4699 struct amdgpu_ring *ring = NULL; local 4734 struct amdgpu_ring *ring; local 4816 struct amdgpu_ring *ring = &adev->gfx.compute_ring[i]; local 5015 struct amdgpu_ring *ring = &adev->gfx.compute_ring[i]; local 5110 struct amdgpu_ring *ring = &adev->gfx.compute_ring[i]; local 5153 gfx_v8_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument 5581 gfx_v8_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned vmid) argument 6010 gfx_v8_0_ring_get_rptr(struct amdgpu_ring *ring) argument 6015 gfx_v8_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument 6026 gfx_v8_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument 6040 gfx_v8_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 6073 gfx_v8_0_ring_emit_vgt_flush(struct amdgpu_ring *ring) argument 6084 gfx_v8_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 6116 gfx_v8_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 6150 gfx_v8_0_ring_emit_fence_gfx(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 6187 gfx_v8_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 6204 gfx_v8_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 6230 gfx_v8_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument 6235 gfx_v8_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument 6244 gfx_v8_0_ring_emit_fence_compute(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 6265 gfx_v8_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 6290 gfx_v8_ring_emit_sb(struct amdgpu_ring *ring) argument 6296 gfx_v8_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument 6329 gfx_v8_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 6345 gfx_v8_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument 6362 gfx_v8_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 6386 gfx_v8_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument 6594 struct amdgpu_ring *ring; local 6624 struct amdgpu_ring *ring; local 6774 gfx_v8_0_emit_mem_sync(struct amdgpu_ring *ring) argument 6787 gfx_v8_0_emit_mem_sync_compute(struct amdgpu_ring *ring) argument 6805 gfx_v8_0_emit_wave_limit_cs(struct amdgpu_ring *ring, uint32_t pipe, bool enable) argument 6836 gfx_v8_0_emit_wave_limit(struct amdgpu_ring *ring, bool enable) argument 7159 gfx_v8_0_ring_emit_ce_meta(struct amdgpu_ring *ring) argument 7188 gfx_v8_0_ring_emit_de_meta(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_vce.c | 85 static int amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, 87 static int amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, 221 amdgpu_ring_fini(&adev->vce.ring[i]); 233 * @ring: amdgpu_ring pointer to check 237 int amdgpu_vce_entity_init(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 239 if (ring == &adev->vce.ring[0]) { 240 struct drm_gpu_scheduler *sched = &ring->sched; 339 count += amdgpu_fence_count_emitted(&adev->vce.ring[i]); 359 * @ring 363 amdgpu_vce_ring_begin_use(struct amdgpu_ring *ring) argument 395 amdgpu_vce_ring_end_use(struct amdgpu_ring *ring) argument 411 struct amdgpu_ring *ring = &adev->vce.ring[0]; local 438 amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct dma_fence **fence) argument 530 amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, bool direct, struct dma_fence **fence) argument 1069 amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 1089 amdgpu_vce_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 1108 amdgpu_vce_ring_test_ring(struct amdgpu_ring *ring) argument 1147 amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1175 amdgpu_vce_get_ring_prio(int ring) argument [all...] |
/linux-master/drivers/net/ethernet/apm/xgene-v2/ |
H A D | main.c | 67 struct xge_desc_ring *ring = pdata->rx_ring; local 72 u8 tail = ring->tail; 79 raw_desc = &ring->raw_desc[tail]; 93 ring->pkt_info[tail].skb = skb; 94 ring->pkt_info[tail].dma_addr = dma_addr; 109 ring->tail = tail; 352 struct xge_desc_ring *ring) 358 if (!ring) 362 if (ring->desc_addr) 363 dma_free_coherent(dev, size, ring 351 xge_delete_desc_ring(struct net_device *ndev, struct xge_desc_ring *ring) argument 372 struct xge_desc_ring *ring = pdata->rx_ring; local 407 struct xge_desc_ring *ring; local 440 struct xge_desc_ring *ring; local [all...] |
/linux-master/drivers/net/can/spi/mcp251xfd/ |
H A D | mcp251xfd-ethtool.c | 16 struct ethtool_ringparam *ring, 25 ring->rx_max_pending = layout.max_rx; 26 ring->tx_max_pending = layout.max_tx; 28 ring->rx_pending = priv->rx_obj_num; 29 ring->tx_pending = priv->tx->obj_num; 34 struct ethtool_ringparam *ring, 42 can_ram_get_layout(&layout, &mcp251xfd_ram_config, ring, NULL, fd_mode); 93 const struct ethtool_ringparam ring = { local 99 can_ram_get_layout(&layout, &mcp251xfd_ram_config, &ring, ec, fd_mode); 15 mcp251xfd_ring_get_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument 33 mcp251xfd_ring_set_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument
|
/linux-master/drivers/net/ethernet/netronome/nfp/flower/ |
H A D | metadata.c | 43 struct circ_buf *ring; local 45 ring = &priv->stats_ids.free_list; 47 if (!CIRC_SPACE(ring->head, ring->tail, priv->stats_ring_size)) 51 memcpy(&ring->buf[ring->head * NFP_FL_STATS_ELEM_RS], 53 ring->head = (ring->head + 1) & (priv->stats_ring_size - 1); 62 struct circ_buf *ring; local 64 ring 140 struct circ_buf *ring; local 165 struct circ_buf *ring; local [all...] |
/linux-master/drivers/net/ethernet/sfc/ |
H A D | ethtool.c | 162 struct ethtool_ringparam *ring, 168 ring->rx_max_pending = EFX_MAX_DMAQ_SIZE; 169 ring->tx_max_pending = EFX_TXQ_MAX_ENT(efx); 170 ring->rx_pending = efx->rxq_entries; 171 ring->tx_pending = efx->txq_entries; 176 struct ethtool_ringparam *ring, 183 if (ring->rx_mini_pending || ring->rx_jumbo_pending || 184 ring->rx_pending > EFX_MAX_DMAQ_SIZE || 185 ring 161 efx_ethtool_get_ringparam(struct net_device *net_dev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument 175 efx_ethtool_set_ringparam(struct net_device *net_dev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/drivers/net/ethernet/sfc/siena/ |
H A D | ethtool.c | 162 struct ethtool_ringparam *ring, 168 ring->rx_max_pending = EFX_MAX_DMAQ_SIZE; 169 ring->tx_max_pending = EFX_TXQ_MAX_ENT(efx); 170 ring->rx_pending = efx->rxq_entries; 171 ring->tx_pending = efx->txq_entries; 176 struct ethtool_ringparam *ring, 183 if (ring->rx_mini_pending || ring->rx_jumbo_pending || 184 ring->rx_pending > EFX_MAX_DMAQ_SIZE || 185 ring 161 efx_ethtool_get_ringparam(struct net_device *net_dev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument 175 efx_ethtool_set_ringparam(struct net_device *net_dev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.h | 325 /* struct that defines a descriptor ring, associated with a VSI */ 327 struct i40e_ring *next; /* pointer to next ring in q_vector */ 328 void *desc; /* Descriptor ring memory */ 330 struct net_device *netdev; /* netdev ring maps to */ 338 u16 queue_index; /* Queue number of ring */ 339 u8 dcb_tc; /* Traffic class of ring */ 342 /* Storing xdp_buff on ring helps in saving the state of partially built 344 * and to resume packet building for this ring in the next call to 361 u16 reg_idx; /* HW register index of the ring */ 372 bool ring_active; /* is ring onlin 404 ring_uses_build_skb(struct i40e_ring *ring) argument 409 set_ring_build_skb_enabled(struct i40e_ring *ring) argument 414 clear_ring_build_skb_enabled(struct i40e_ring *ring) argument 419 ring_is_xdp(struct i40e_ring *ring) argument 424 set_ring_xdp(struct i40e_ring *ring) argument 436 struct i40e_ring *ring; /* pointer to linked list of ring(s) */ member in struct:i40e_ring_container 449 i40e_rx_pg_order(struct i40e_ring *ring) argument 561 txring_txq(const struct i40e_ring *ring) argument [all...] |
/linux-master/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_cq.h | 49 struct vnic_dev_ring ring; member in struct:vnic_cq 70 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 71 cq->ring.desc_size * cq->to_clean); 82 if (cq->to_clean == cq->ring.desc_count) { 87 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 88 cq->ring.desc_size * cq->to_clean);
|
/linux-master/drivers/scsi/fnic/ |
H A D | vnic_cq.h | 53 struct vnic_dev_ring ring; member in struct:vnic_cq 69 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 70 cq->ring.desc_size * cq->to_clean); 81 if (cq->to_clean == cq->ring.desc_count) { 86 cq_desc = (struct cq_desc *)((u8 *)cq->ring.descs + 87 cq->ring.desc_size * cq->to_clean);
|
/linux-master/drivers/net/ethernet/atheros/ |
H A D | ag71xx.c | 393 static struct ag71xx_desc *ag71xx_ring_desc(struct ag71xx_ring *ring, int idx) argument 395 return (struct ag71xx_desc *)&ring->descs_cpu[idx * AG71XX_DESC_SIZE]; 792 struct ag71xx_ring *ring = &ag->tx_ring; local 798 ring_mask = BIT(ring->order) - 1; 799 ring_size = BIT(ring->order); 801 netif_dbg(ag, tx_queued, ndev, "processing TX ring\n"); 803 while (ring->dirty + n != ring->curr) { 808 i = (ring->dirty + n) & ring_mask; 809 desc = ag71xx_ring_desc(ring, 1141 struct ag71xx_ring *ring = &ag->tx_ring; local 1173 struct ag71xx_ring *ring = &ag->tx_ring; local 1198 struct ag71xx_ring *ring = &ag->rx_ring; local 1224 struct ag71xx_ring *ring = &ag->rx_ring; local 1243 struct ag71xx_ring *ring = &ag->rx_ring; local 1284 struct ag71xx_ring *ring = &ag->rx_ring; local 1466 ag71xx_fill_dma_desc(struct ag71xx_ring *ring, u32 addr, int len) argument 1520 struct ag71xx_ring *ring; local 1621 struct ag71xx_ring *ring; local [all...] |
/linux-master/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
H A D | hw_atl_a0.c | 406 struct aq_ring_s *ring) 408 hw_atl_tdm_tx_desc_en_set(self, 1, ring->idx); 414 struct aq_ring_s *ring) 416 hw_atl_rdm_rx_desc_en_set(self, 1, ring->idx); 430 struct aq_ring_s *ring) 432 hw_atl_reg_tx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); 438 struct aq_ring_s *ring, 448 buff = &ring->buff_ring[ring 405 hw_atl_a0_hw_ring_tx_start(struct aq_hw_s *self, struct aq_ring_s *ring) argument 413 hw_atl_a0_hw_ring_rx_start(struct aq_hw_s *self, struct aq_ring_s *ring) argument 429 hw_atl_a0_hw_tx_ring_tail_update(struct aq_hw_s *self, struct aq_ring_s *ring) argument 437 hw_atl_a0_hw_ring_tx_xmit(struct aq_hw_s *self, struct aq_ring_s *ring, unsigned int frags) argument 585 hw_atl_a0_hw_ring_rx_fill(struct aq_hw_s *self, struct aq_ring_s *ring, unsigned int sw_tail_old) argument 606 hw_atl_a0_hw_ring_tx_head_update(struct aq_hw_s *self, struct aq_ring_s *ring) argument 623 hw_atl_a0_hw_ring_rx_receive(struct aq_hw_s *self, struct aq_ring_s *ring) argument 874 hw_atl_a0_hw_ring_tx_stop(struct aq_hw_s *self, struct aq_ring_s *ring) argument 882 hw_atl_a0_hw_ring_rx_stop(struct aq_hw_s *self, struct aq_ring_s *ring) argument [all...] |
/linux-master/drivers/bus/mhi/ep/ |
H A D | main.c | 33 struct mhi_ep_ring *ring; local 37 ring = &mhi_cntrl->mhi_event[ring_idx].ring; 39 if (!ring->started) { 40 ret = mhi_ep_ring_start(mhi_cntrl, ring, ctx); 42 dev_err(dev, "Error starting event ring (%u)\n", ring_idx); 47 /* Add element to the event ring */ 48 ret = mhi_ep_ring_add_element(ring, el); 50 dev_err(dev, "Error adding element to event ring (%u)\n", ring_idx); 60 * for the event ring an 87 mhi_ep_send_completion_event(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_ring *ring, struct mhi_ring_element *tre, u32 len, enum mhi_ev_ccs code) argument 145 struct mhi_ep_ring *ring = &mhi_cntrl->mhi_cmd->ring; local 163 mhi_ep_process_cmd_ring(struct mhi_ep_ring *ring, struct mhi_ring_element *el) argument 331 struct mhi_ep_ring *ring = &mhi_cntrl->mhi_chan[mhi_chan->chan].ring; local 342 struct mhi_ep_ring *ring = &mhi_cntrl->mhi_chan[mhi_chan->chan].ring; local 401 mhi_ep_read_channel(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_ring *ring) argument 481 mhi_ep_process_ch_ring(struct mhi_ep_ring *ring) argument 524 struct mhi_ep_ring *ring = &mhi_cntrl->mhi_chan[mhi_chan->chan].ring; local 557 struct mhi_ep_ring *ring; local 783 struct mhi_ep_ring *ring = &mhi_cntrl->mhi_cmd->ring; local 819 struct mhi_ep_ring *ring; local 919 struct mhi_ep_ring *ring; local [all...] |
/linux-master/drivers/net/thunderbolt/ |
H A D | main.c | 58 * Each data frame passed to the high-speed DMA ring has this header. If 145 struct tb_ring *ring; member in struct:tbnet_ring 175 * @rx_ring: Software ring holding Rx frames 178 * @tx_ring: Software ring holding Tx frames 334 static void tbnet_free_buffers(struct tbnet_ring *ring) argument 339 struct device *dma_dev = tb_ring_dma_device(ring->ring); 340 struct tbnet_frame *tf = &ring->frames[i]; 348 if (ring->ring 494 tbnet_available_buffers(const struct tbnet_ring *ring) argument 501 struct tbnet_ring *ring = &net->rx_ring; local 550 struct tbnet_ring *ring = &net->tx_ring; local 569 tbnet_tx_callback(struct tb_ring *ring, struct ring_frame *frame, bool canceled) argument 584 struct tbnet_ring *ring = &net->tx_ring; local 921 struct tb_ring *ring; local [all...] |
/linux-master/drivers/bus/mhi/host/ |
H A D | init.c | 162 /* MHI protocol requires the transfer ring to be aligned with ring length */ 164 struct mhi_ring *ring, 167 ring->alloc_size = len + (len - 1); 168 ring->pre_aligned = dma_alloc_coherent(mhi_cntrl->cntrl_dev, ring->alloc_size, 169 &ring->dma_handle, GFP_KERNEL); 170 if (!ring->pre_aligned) 173 ring->iommu_base = (ring 163 mhi_alloc_aligned_ring(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring, u64 len) argument 263 struct mhi_ring *ring; local 366 struct mhi_ring *ring = &mhi_event->ring; local 412 struct mhi_ring *ring = &mhi_cmd->ring; local 434 struct mhi_ring *ring = &mhi_cmd->ring; local 447 struct mhi_ring *ring = &mhi_event->ring; local [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_fwlog.c | 45 struct ice_fwlog_data *ring = &rings->rings[i]; local 47 ring->data_size = ICE_AQ_MAX_BUF_LEN; 48 ring->data = mem; 60 struct ice_fwlog_data *ring = &rings->rings[i]; local 62 /* the first ring is the base memory for the whole range so 66 vfree(ring->data); 68 ring->data = NULL; 69 ring->data_size = 0; 82 struct ice_fwlog_ring ring; local 88 * driver uses a ring o [all...] |
/linux-master/drivers/net/can/c_can/ |
H A D | c_can.h | 179 /* c_can tx ring structure */ 228 static inline u8 c_can_get_tx_head(const struct c_can_tx_ring *ring) argument 230 return ring->head & (ring->obj_num - 1); 233 static inline u8 c_can_get_tx_tail(const struct c_can_tx_ring *ring) argument 235 return ring->tail & (ring->obj_num - 1); 239 const struct c_can_tx_ring *ring) 241 u8 head = c_can_get_tx_head(ring); 242 u8 tail = c_can_get_tx_tail(ring); 238 c_can_get_tx_free(const struct c_can_priv *priv, const struct c_can_tx_ring *ring) argument [all...] |