/linux-master/drivers/net/ethernet/broadcom/genet/ |
H A D | bcmgenet.c | 345 /* RDMA/TDMA ring registers and accessors 409 unsigned int ring, 413 (DMA_RING_SIZE * ring) + 418 unsigned int ring, u32 val, 422 (DMA_RING_SIZE * ring) + 427 unsigned int ring, 431 (DMA_RING_SIZE * ring) + 436 unsigned int ring, u32 val, 440 (DMA_RING_SIZE * ring) + 672 * We'll map them to ring 408 bcmgenet_tdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument 417 bcmgenet_tdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument 426 bcmgenet_rdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument 435 bcmgenet_rdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument 818 struct bcmgenet_rx_ring *ring; local 840 bcmgenet_set_rx_coalesce(struct bcmgenet_rx_ring *ring, u32 usecs, u32 pkts) argument 855 bcmgenet_set_ring_rx_coalesce(struct bcmgenet_rx_ring *ring, struct ethtool_coalesce *ec) argument 1728 bcmgenet_get_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument 1745 bcmgenet_put_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument 1762 bcmgenet_rx_ring16_int_disable(struct bcmgenet_rx_ring *ring) argument 1768 bcmgenet_rx_ring16_int_enable(struct bcmgenet_rx_ring *ring) argument 1774 bcmgenet_rx_ring_int_disable(struct bcmgenet_rx_ring *ring) argument 1781 bcmgenet_rx_ring_int_enable(struct bcmgenet_rx_ring *ring) argument 1788 bcmgenet_tx_ring16_int_disable(struct bcmgenet_tx_ring *ring) argument 1794 bcmgenet_tx_ring16_int_enable(struct bcmgenet_tx_ring *ring) argument 1800 bcmgenet_tx_ring_int_enable(struct bcmgenet_tx_ring *ring) argument 1806 bcmgenet_tx_ring_int_disable(struct bcmgenet_tx_ring *ring) argument 1868 __bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument 1925 bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument 1939 struct bcmgenet_tx_ring *ring = local 2047 struct bcmgenet_tx_ring *ring = NULL; local 2229 bcmgenet_desc_rx(struct bcmgenet_rx_ring *ring, unsigned int budget) argument 2393 struct bcmgenet_rx_ring *ring = container_of(napi, local 2419 struct bcmgenet_rx_ring *ring = local 2429 bcmgenet_alloc_rx_buffers(struct bcmgenet_priv *priv, struct bcmgenet_rx_ring *ring) argument 2599 bcmgenet_init_dim(struct bcmgenet_rx_ring *ring, void (*cb)(struct work_struct *work)) argument 2611 bcmgenet_init_rx_coalesce(struct bcmgenet_rx_ring *ring) argument 2635 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; local 2694 struct bcmgenet_rx_ring *ring = &priv->rx_rings[index]; local 2750 struct bcmgenet_tx_ring *ring; local 2766 struct bcmgenet_tx_ring *ring; local 2780 struct bcmgenet_tx_ring *ring; local 2863 struct bcmgenet_rx_ring *ring; local 2879 struct bcmgenet_rx_ring *ring; local 2895 struct bcmgenet_rx_ring *ring; local 3498 bcmgenet_dump_tx_queue(struct bcmgenet_tx_ring *ring) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 170 * @ring: DMA ring to use for the copy 181 unsigned int window, struct amdgpu_ring *ring, 184 struct amdgpu_device *adev = ring->adev; 241 amdgpu_ring_pad_ib(ring, &job->ibs[0]); 294 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local 300 DRM_ERROR("Trying to move memory with ring turned off.\n"); 317 0, ring, tmz, &cur_size, &from); 322 1, ring, tmz, &cur_size, &to); 326 r = amdgpu_copy_buffer(ring, fro 178 amdgpu_ttm_map_buffer(struct ttm_buffer_object *bo, struct ttm_resource *mem, struct amdgpu_res_cursor *mm_cur, unsigned int window, struct amdgpu_ring *ring, bool tmz, uint64_t *size, uint64_t *addr) argument 2066 struct amdgpu_ring *ring; local 2142 amdgpu_copy_buffer(struct amdgpu_ring *ring, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, struct dma_resv *resv, struct dma_fence **fence, bool direct_submit, bool vm_needs_flush, bool tmz) argument 2196 amdgpu_ttm_fill_mem(struct amdgpu_ring *ring, uint32_t src_data, uint64_t dst_addr, uint32_t byte_count, struct dma_resv *resv, struct dma_fence **fence, bool vm_needs_flush, bool delayed) argument 2240 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local [all...] |
H A D | amdgpu_object.c | 754 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local 760 return amdgpu_copy_buffer(ring, shadow_addr, parent_addr,
|
H A D | vpe_v6_1.c | 67 base = vpe->ring.adev->reg_offset[VPE_HWIP][inst][0]; 74 struct amdgpu_device *adev = vpe->ring.adev; 101 struct amdgpu_device *adev = vpe->ring.adev; 124 struct amdgpu_device *adev = vpe->ring.adev; 207 struct amdgpu_ring *ring = &vpe->ring; local 208 struct amdgpu_device *adev = ring->adev; 215 /* Set ring buffer size in dwords */ 216 rb_bufsz = order_base_2(ring->ring_size / 4); 223 /* Initialize the ring buffe [all...] |
H A D | sdma_v5_2.c | 92 static unsigned sdma_v5_2_ring_init_cond_exec(struct amdgpu_ring *ring, argument 97 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE)); 98 amdgpu_ring_write(ring, lower_32_bits(addr)); 99 amdgpu_ring_write(ring, upper_32_bits(addr)); 100 amdgpu_ring_write(ring, 1); 102 ret = ring->wptr & ring->buf_mask; 104 amdgpu_ring_write(ring, 0); 112 * @ring: amdgpu ring pointe 116 sdma_v5_2_ring_get_rptr(struct amdgpu_ring *ring) argument 134 sdma_v5_2_ring_get_wptr(struct amdgpu_ring *ring) argument 160 sdma_v5_2_ring_set_wptr(struct amdgpu_ring *ring) argument 194 sdma_v5_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 217 sdma_v5_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 252 sdma_v5_2_ring_emit_mem_sync(struct amdgpu_ring *ring) argument 277 sdma_v5_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 312 sdma_v5_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 475 struct amdgpu_ring *ring; local 809 sdma_v5_2_ring_test_ring(struct amdgpu_ring *ring) argument 885 sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1066 sdma_v5_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1091 sdma_v5_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1120 sdma_v5_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1126 sdma_v5_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1135 sdma_v5_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1149 sdma_v5_2_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 1213 struct amdgpu_ring *ring; local 1331 sdma_v5_2_ring_preempt_ib(struct amdgpu_ring *ring) argument 1638 sdma_v5_2_ring_begin_use(struct amdgpu_ring *ring) argument 1653 sdma_v5_2_ring_end_use(struct amdgpu_ring *ring) argument [all...] |
H A D | sdma_v4_4_2.c | 154 * @ring: amdgpu ring pointer 158 static uint64_t sdma_v4_4_2_ring_get_rptr(struct amdgpu_ring *ring) argument 163 rptr = READ_ONCE(*((u64 *)&ring->adev->wb.wb[ring->rptr_offs])); 172 * @ring: amdgpu ring pointer 176 static uint64_t sdma_v4_4_2_ring_get_wptr(struct amdgpu_ring *ring) argument 178 struct amdgpu_device *adev = ring->adev; 181 if (ring 203 sdma_v4_4_2_ring_set_wptr(struct amdgpu_ring *ring) argument 245 sdma_v4_4_2_page_ring_get_wptr(struct amdgpu_ring *ring) argument 269 sdma_v4_4_2_page_ring_set_wptr(struct amdgpu_ring *ring) argument 289 sdma_v4_4_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 312 sdma_v4_4_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 333 sdma_v4_4_2_wait_reg_mem(struct amdgpu_ring *ring, int mem_space, int hdp, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 365 sdma_v4_4_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 392 sdma_v4_4_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 596 sdma_v4_4_2_rb_cntl(struct amdgpu_ring *ring, uint32_t rb_cntl) argument 622 struct amdgpu_ring *ring = &adev->sdma.instance[i].ring; local 710 struct amdgpu_ring *ring = &adev->sdma.instance[i].page; local 867 struct amdgpu_ring *ring; local 954 sdma_v4_4_2_ring_test_ring(struct amdgpu_ring *ring) argument 1007 sdma_v4_4_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1159 sdma_v4_4_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1184 sdma_v4_4_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1207 sdma_v4_4_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1213 sdma_v4_4_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1222 sdma_v4_4_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1282 struct amdgpu_ring *ring; local [all...] |
H A D | gfx_v9_0.c | 758 static void gfx_v9_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume, bool usegds); 759 static u64 gfx_v9_0_ring_get_rptr_compute(struct amdgpu_ring *ring); 787 struct amdgpu_ring *ring) 789 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj); 790 uint64_t wptr_addr = ring->wptr_gpu_addr; 791 uint32_t eng_sel = ring->funcs->type == AMDGPU_RING_TYPE_GFX ? 4 : 0; 798 PACKET3_MAP_QUEUES_QUEUE(ring->queue) | 799 PACKET3_MAP_QUEUES_PIPE(ring->pipe) | 800 PACKET3_MAP_QUEUES_ME((ring->me == 1 ? 0 : 1)) | 809 PACKET3_MAP_QUEUES_DOORBELL_OFFSET(ring 786 gfx_v9_0_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument 816 gfx_v9_0_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument 844 gfx_v9_0_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument 960 gfx_v9_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument 972 gfx_v9_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 994 gfx_v9_0_ring_test_ring(struct amdgpu_ring *ring) argument 1024 gfx_v9_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1969 struct amdgpu_ring *ring = &adev->gfx.compute_ring[ring_id]; local 2001 struct amdgpu_ring *ring; local 3025 struct amdgpu_ring *ring = &adev->gfx.gfx_ring[0]; local 3095 struct amdgpu_ring *ring; local 3217 gfx_v9_0_kiq_setting(struct amdgpu_ring *ring) argument 3231 gfx_v9_0_mqd_set_priority(struct amdgpu_ring *ring, struct v9_mqd *mqd) argument 3244 gfx_v9_0_mqd_init(struct amdgpu_ring *ring) argument 3376 gfx_v9_0_kiq_init_register(struct amdgpu_ring *ring) argument 3490 gfx_v9_0_kiq_fini_register(struct amdgpu_ring *ring) argument 3529 gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring) argument 3577 gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring) argument 3617 struct amdgpu_ring *ring; local 3641 struct amdgpu_ring *ring = NULL; local 3671 struct amdgpu_ring *ring; local 3926 struct amdgpu_ring *ring = &kiq->ring; local 4032 gfx_v9_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument 4299 struct amdgpu_ring *ring = &adev->gfx.compute_ring[0]; local 4346 struct amdgpu_ring *ring = &adev->gfx.compute_ring[0]; local 4904 gfx_v9_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned int vmid) argument 5073 gfx_v9_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument 5078 gfx_v9_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument 5094 gfx_v9_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument 5108 gfx_v9_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 5137 gfx_v9_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5178 gfx_v9_0_ring_patch_cntl(struct amdgpu_ring *ring, unsigned offset) argument 5187 gfx_v9_0_ring_patch_ce_meta(struct amdgpu_ring *ring, unsigned offset) argument 5219 gfx_v9_0_ring_patch_de_meta(struct amdgpu_ring *ring, unsigned offset) argument 5254 gfx_v9_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5289 gfx_v9_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 5330 gfx_v9_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 5341 gfx_v9_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 5354 gfx_v9_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument 5359 gfx_v9_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument 5371 gfx_v9_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument 5384 gfx_v9_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 5411 gfx_v9_ring_emit_sb(struct amdgpu_ring *ring) argument 5417 gfx_v9_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume) argument 5459 gfx_v9_0_ring_preempt_ib(struct amdgpu_ring *ring) argument 5517 gfx_v9_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume, bool usegds) argument 5571 gfx_v9_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument 5580 gfx_v9_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument 5613 gfx_v9_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 5628 gfx_v9_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument 5645 gfx_v9_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 5668 gfx_v9_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 5674 gfx_v9_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 5691 gfx_v9_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument 5893 struct amdgpu_ring *ring; local 5931 struct amdgpu_ring *ring; local 6765 gfx_v9_0_emit_mem_sync(struct amdgpu_ring *ring) argument 6784 gfx_v9_0_emit_wave_limit_cs(struct amdgpu_ring *ring, uint32_t pipe, bool enable) argument 6815 gfx_v9_0_emit_wave_limit(struct amdgpu_ring *ring, bool enable) argument [all...] |
H A D | gfx_v11_0.c | 133 static void gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume); 134 static void gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure); 135 static void gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, 138 static void gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, 161 struct amdgpu_ring *ring) 163 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj); 164 uint64_t wptr_addr = ring->wptr_gpu_addr; 167 switch (ring->funcs->type) { 189 PACKET3_MAP_QUEUES_QUEUE(ring->queue) | 190 PACKET3_MAP_QUEUES_PIPE(ring 160 gfx11_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument 203 gfx11_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument 236 gfx11_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument 303 gfx_v11_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument 314 gfx_v11_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 336 gfx_v11_0_ring_test_ring(struct amdgpu_ring *ring) argument 377 gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 933 struct amdgpu_ring *ring; local 965 struct amdgpu_ring *ring; local 3151 struct amdgpu_ring *ring; local 3236 gfx_v11_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 3263 struct amdgpu_ring *ring; local 3589 gfx_v11_0_kiq_setting(struct amdgpu_ring *ring) argument 3704 gfx_v11_0_gfx_init_queue(struct amdgpu_ring *ring) argument 3735 struct amdgpu_ring *ring; local 3892 gfx_v11_0_kiq_init_register(struct amdgpu_ring *ring) argument 4002 gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring) argument 4041 gfx_v11_0_kcq_init_queue(struct amdgpu_ring *ring) argument 4072 struct amdgpu_ring *ring; local 4097 struct amdgpu_ring *ring = NULL; local 4128 struct amdgpu_ring *ring; local 4630 struct amdgpu_ring *ring; local 4687 gfx_v11_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument 5036 gfx_v11_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned vmid) argument 5211 gfx_v11_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument 5217 gfx_v11_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument 5233 gfx_v11_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument 5250 gfx_v11_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument 5256 gfx_v11_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument 5268 gfx_v11_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument 5282 gfx_v11_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 5311 gfx_v11_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5351 gfx_v11_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5390 gfx_v11_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 5428 gfx_v11_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 5438 gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument 5450 gfx_v11_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 5472 gfx_v11_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 5499 gfx_v11_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument 5519 gfx_v11_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 5536 gfx_v11_0_ring_emit_gfx_shadow(struct amdgpu_ring *ring, u64 shadow_va, u64 csa_va, u64 gds_va, bool init_shadow, int vmid) argument 5627 gfx_v11_0_ring_preempt_ib(struct amdgpu_ring *ring) argument 5674 gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument 5725 gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument 5734 gfx_v11_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument 5751 gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 5774 gfx_v11_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 5780 gfx_v11_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 5790 gfx_v11_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument 5942 struct amdgpu_ring *ring; local 6033 struct amdgpu_ring *ring; local 6136 gfx_v11_0_emit_mem_sync(struct amdgpu_ring *ring) argument [all...] |
H A D | gfx_v10_0.c | 52 * 1. Primary ring 53 * 2. Async ring 3489 static void gfx_v10_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume); 3490 static void gfx_v10_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume); 3491 static void gfx_v10_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure); 3495 static void gfx_v10_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, 3517 struct amdgpu_ring *ring) 3519 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj); 3520 uint64_t wptr_addr = ring->wptr_gpu_addr; 3523 switch (ring 3516 gfx10_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument 3556 gfx10_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument 3589 gfx10_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument 3743 gfx_v10_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument 3754 gfx_v10_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 3776 gfx_v10_0_ring_test_ring(struct amdgpu_ring *ring) argument 3814 gfx_v10_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 4434 struct amdgpu_ring *ring; local 4465 struct amdgpu_ring *ring; local 5961 struct amdgpu_ring *ring; local 6050 gfx_v10_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 6097 struct amdgpu_ring *ring; local 6311 gfx_v10_0_kiq_setting(struct amdgpu_ring *ring) argument 6445 gfx_v10_0_gfx_init_queue(struct amdgpu_ring *ring) argument 6492 struct amdgpu_ring *ring; local 6635 gfx_v10_0_kiq_init_register(struct amdgpu_ring *ring) argument 6744 gfx_v10_0_kiq_init_queue(struct amdgpu_ring *ring) argument 6783 gfx_v10_0_kcq_init_queue(struct amdgpu_ring *ring) argument 6814 struct amdgpu_ring *ring; local 6838 struct amdgpu_ring *ring = NULL; local 6868 struct amdgpu_ring *ring; local 7406 gfx_v10_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument 7953 gfx_v10_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned int vmid) argument 8174 gfx_v10_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument 8180 gfx_v10_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument 8196 gfx_v10_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument 8241 gfx_v10_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument 8247 gfx_v10_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument 8259 gfx_v10_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument 8301 gfx_v10_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 8330 gfx_v10_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 8371 gfx_v10_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 8410 gfx_v10_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 8444 gfx_v10_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 8454 gfx_v10_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument 8466 gfx_v10_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument 8482 gfx_v10_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 8509 gfx_v10_0_ring_emit_sb(struct amdgpu_ring *ring) argument 8515 gfx_v10_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument 8549 gfx_v10_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 8566 gfx_v10_0_ring_preempt_ib(struct amdgpu_ring *ring) argument 8613 gfx_v10_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume) argument 8653 gfx_v10_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument 8704 gfx_v10_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument 8713 gfx_v10_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument 8730 gfx_v10_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 8753 gfx_v10_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 8759 gfx_v10_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 8777 gfx_v10_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned int vmid) argument 8933 struct amdgpu_ring *ring; local 9024 struct amdgpu_ring *ring; local 9078 struct amdgpu_ring *ring = &(adev->gfx.kiq[0].ring); local 9122 struct amdgpu_ring *ring = &(adev->gfx.kiq[0].ring); local 9134 gfx_v10_0_emit_mem_sync(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_vpe.c | 123 struct amdgpu_device *adev = vpe->ring.adev; 224 struct amdgpu_device *adev = vpe->ring.adev; 267 struct amdgpu_ring *ring = &vpe->ring; local 270 ring->ring_obj = NULL; 271 ring->use_doorbell = true; 272 ring->vm_hub = AMDGPU_MMHUB0(0); 273 ring->doorbell_index = (adev->doorbell_index.vpe_ring << 1); 274 snprintf(ring->name, 4, "vpe"); 276 ret = amdgpu_ring_init(adev, ring, 102 445 vpe_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 457 vpe_get_csa_mc_addr(struct amdgpu_ring *ring, uint32_t vmid) argument 472 vpe_ring_emit_pred_exec(struct amdgpu_ring *ring, uint32_t device_select, uint32_t exec_count) argument 484 vpe_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 503 vpe_ring_emit_fence(struct amdgpu_ring *ring, uint64_t addr, uint64_t seq, unsigned int flags) argument 527 vpe_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 547 vpe_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 556 vpe_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 573 vpe_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument 579 vpe_ring_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 594 vpe_ring_preempt_ib(struct amdgpu_ring *ring) argument 662 vpe_ring_get_rptr(struct amdgpu_ring *ring) argument 681 vpe_ring_get_wptr(struct amdgpu_ring *ring) argument 700 vpe_ring_set_wptr(struct amdgpu_ring *ring) argument 734 vpe_ring_test_ring(struct amdgpu_ring *ring) argument 776 vpe_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 830 vpe_ring_begin_use(struct amdgpu_ring *ring) argument 855 vpe_ring_end_use(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_umsch_mm.c | 269 uint32_t *ring = test->ring_data_cpu_addr + local 285 ring[0] = VPE_CMD_HEADER(VPE_CMD_OPCODE_INDIRECT, 0); 286 ring[1] = (ib_gpu_addr & 0xffffffe0); 287 ring[2] = upper_32_bits(ib_gpu_addr); 288 ring[3] = 4; 289 ring[4] = 0; 290 ring[5] = 0; 491 struct amdgpu_ring *ring = &umsch->ring; local 493 if (amdgpu_ring_alloc(ring, ndw 504 struct amdgpu_ring *ring = &umsch->ring; local 518 umsch_mm_ring_set_wptr(struct amdgpu_ring *ring) argument 529 umsch_mm_ring_get_rptr(struct amdgpu_ring *ring) argument 537 umsch_mm_ring_get_wptr(struct amdgpu_ring *ring) argument 559 struct amdgpu_ring *ring = &umsch->ring; local [all...] |
H A D | amdgpu_mes.c | 171 "(%d) ring trail_fence_offs wb alloc failed\n", r); 596 amdgpu_gfx_select_me_pipe_q(adev, p->ring->me, p->ring->pipe, 0, 0, 0); 704 queue->ring = qprops->ring; 710 queue->ring->mqd_ptr = queue->mqd_cpu_ptr; 782 struct amdgpu_ring *ring, 790 queue_input.queue_type = ring->funcs->type; 791 queue_input.doorbell_offset = ring->doorbell_index; 792 queue_input.pipe_id = ring 781 amdgpu_mes_unmap_legacy_queue(struct amdgpu_device *adev, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument 973 amdgpu_mes_ring_to_queue_props(struct amdgpu_device *adev, struct amdgpu_ring *ring, struct amdgpu_mes_queue_properties *props) argument 1007 amdgpu_mes_ctx_get_offs(struct amdgpu_ring *ring, unsigned int id_offs) argument 1032 struct amdgpu_ring *ring; local 1128 amdgpu_mes_remove_ring(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 1310 struct amdgpu_ring *ring; local 1345 struct amdgpu_ring *ring; local [all...] |
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 228 /* the page has been released from the ring */ 326 * @rx_ring: ring to place buffers on 375 /* update next to alloc since we have filled the ring */ 388 u16 wx_desc_unused(struct wx_ring *ring) argument 390 u16 ntc = ring->next_to_clean; 391 u16 ntu = ring->next_to_use; 393 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; 398 * @rx_ring: Rx ring being processed 458 * @rx_ring: rx descriptor ring packet is being transacted on 499 static void wx_rx_hash(struct wx_ring *ring, argument 525 wx_rx_checksum(struct wx_ring *ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument 571 wx_rx_vlan(struct wx_ring *ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument 694 wx_txring_txq(const struct wx_ring *ring) argument 837 struct wx_ring *ring; local 1711 wx_add_ring(struct wx_ring *ring, struct wx_ring_container *head) argument 1737 struct wx_ring *ring; local 1843 struct wx_ring *ring; local 2158 struct wx_ring *ring; local 2637 struct wx_ring *ring = READ_ONCE(wx->rx_ring[i]); local 2653 struct wx_ring *ring = READ_ONCE(wx->tx_ring[i]); local [all...] |
/linux-master/drivers/net/ethernet/renesas/ |
H A D | ravb_main.c | 330 /* Build RX ring buffer */ 370 /* Build TX ring buffer */ 856 /* Refill the RX ring buffers. */ 967 /* Refill the RX ring buffers. */ 1624 struct ethtool_ringparam *ring, 1630 ring->rx_max_pending = BE_RX_RING_MAX; 1631 ring->tx_max_pending = BE_TX_RING_MAX; 1632 ring->rx_pending = priv->num_rx_ring[RAVB_BE]; 1633 ring->tx_pending = priv->num_tx_ring[RAVB_BE]; 1637 struct ethtool_ringparam *ring, 1623 ravb_get_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument 1636 ravb_set_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_main.c | 192 * @tx_ring: ring to be cleaned 257 /* Zero out the buffer ring */ 261 /* Zero out the descriptor ring */ 271 * @tx_ring: Tx descriptor ring for a specific queue 319 static void igc_disable_tx_ring_hw(struct igc_ring *ring) argument 321 struct igc_hw *hw = &ring->q_vector->adapter->hw; 322 u8 idx = ring->reg_idx; 348 * @tx_ring: tx descriptor ring (for a specific queue) to setup 380 netdev_err(ndev, "Unable to allocate memory for Tx descriptor ring\n"); 415 /* Free all the Rx ring sk_buff 443 igc_clean_rx_ring_xsk_pool(struct igc_ring *ring) argument 462 igc_clean_rx_ring(struct igc_ring *ring) argument 608 igc_get_xsk_pool(struct igc_adapter *adapter, struct igc_ring *ring) argument 625 igc_configure_rx_ring(struct igc_adapter *adapter, struct igc_ring *ring) argument 728 igc_configure_tx_ring(struct igc_adapter *adapter, struct igc_ring *ring) argument 1038 igc_tx_launchtime(struct igc_ring *ring, ktime_t txtime, bool *first_flag, bool *insert_empty) argument 1084 igc_init_empty_frame(struct igc_ring *ring, struct igc_tx_buffer *buffer, struct sk_buff *skb) argument 1110 igc_init_tx_empty_descriptor(struct igc_ring *ring, struct sk_buff *skb, struct igc_tx_buffer *first) argument 1715 igc_rx_checksum(struct igc_ring *ring, union igc_adv_rx_desc *rx_desc, struct sk_buff *skb) argument 1775 igc_rx_hash(struct igc_ring *ring, union igc_adv_rx_desc *rx_desc, struct sk_buff *skb) argument 1892 igc_get_rx_frame_truesize(struct igc_ring *ring, unsigned int size) argument 2292 igc_alloc_rx_buffers_zc(struct igc_ring *ring, u16 count) argument 2352 igc_xdp_init_tx_descriptor(struct igc_ring *ring, struct xdp_frame *xdpf) argument 2468 struct igc_ring *ring; local 2535 igc_flush_tx_descriptors(struct igc_ring *ring) argument 2549 struct igc_ring *ring; local 2567 struct igc_ring *ring = q_vector->rx.ring; local 2706 igc_construct_skb_zc(struct igc_ring *ring, struct xdp_buff *xdp) argument 2736 struct igc_ring *ring = q_vector->rx.ring; local 2767 struct igc_ring *ring = q_vector->rx.ring; local 2866 struct igc_ring *ring = q_vector->tx.ring; local 2877 igc_xdp_xmit_zc(struct igc_ring *ring) argument 3944 struct igc_ring *ring = adapter->rx_ring[i]; local 4525 igc_add_ring(struct igc_ring *ring, struct igc_ring_container *head) argument 4612 struct igc_ring *ring; local 4909 struct igc_ring *ring = adapter->rx_ring[i]; local 4935 struct igc_ring *ring = adapter->tx_ring[i]; local 6048 struct igc_ring *ring; local 6149 struct igc_ring *ring = adapter->tx_ring[i]; local 6161 struct igc_ring *ring = adapter->tx_ring[i]; local 6261 struct igc_ring *ring = adapter->tx_ring[i]; local 6289 struct igc_ring *ring = adapter->tx_ring[i]; local 6312 struct igc_ring *ring = adapter->tx_ring[i]; local 6346 struct igc_ring *ring; local 6476 struct igc_ring *ring; local 6526 struct igc_ring *ring; local 7393 igc_disable_rx_ring_hw(struct igc_ring *ring) argument 7405 igc_disable_rx_ring(struct igc_ring *ring) argument 7411 igc_enable_rx_ring(struct igc_ring *ring) argument 7423 igc_disable_tx_ring(struct igc_ring *ring) argument 7429 igc_enable_tx_ring(struct igc_ring *ring) argument [all...] |
H A D | igc.h | 100 struct igc_ring *ring; /* pointer to linked list of rings */ member in struct:igc_ring_container 105 u8 itr; /* current ITR setting for ring */ 116 void *desc; /* descriptor ring memory */ 117 unsigned long flags; /* ring specific flags */ 118 void __iomem *tail; /* pointer to ring tail register */ 119 dma_addr_t dma; /* phys address of the ring */ 120 unsigned int size; /* length of desc. ring in bytes */ 122 u16 count; /* number of desc. in the ring */ 123 u8 queue_index; /* logical index of the ring*/ 124 u8 reg_idx; /* physical index of the ring */ 630 igc_desc_unused(const struct igc_ring *ring) argument 680 igc_rx_bufsz(struct igc_ring *ring) argument 692 igc_rx_pg_order(struct igc_ring *ring) argument [all...] |
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_main.c | 433 if (!q_vector->tx.ring && !q_vector->rx.ring) 455 rx_ring->next = q_vector->rx.ring; 457 q_vector->rx.ring = rx_ring; 481 tx_ring->next = q_vector->tx.ring; 483 q_vector->tx.ring = tx_ring; 499 * one vector per ring/queue, but on a constrained vector budget, we 577 if (q_vector->tx.ring && q_vector->rx.ring) { 581 } else if (q_vector->rx.ring) { 1246 struct iavf_ring *ring = &adapter->rx_rings[i]; local [all...] |
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 432 * i40e_get_netdev_stats_struct_tx - populate stats from a Tx ring 433 * @ring: Tx ring to get statistics from 436 static void i40e_get_netdev_stats_struct_tx(struct i40e_ring *ring, argument 443 start = u64_stats_fetch_begin(&ring->syncp); 444 packets = ring->stats.packets; 445 bytes = ring->stats.bytes; 446 } while (u64_stats_fetch_retry(&ring->syncp, start)); 466 struct i40e_ring *ring; local 480 ring 3415 i40e_config_xps_tx_ring(struct i40e_ring *ring) argument 3437 i40e_xsk_pool(struct i40e_ring *ring) argument 3457 i40e_configure_tx_ring(struct i40e_ring *ring) argument 3581 i40e_configure_rx_ring(struct i40e_ring *ring) argument 5109 struct i40e_ring *ring; local 11639 struct i40e_ring *ring; local [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 398 netdev_err(bp->dev, "Invalid Tx completion (ring:%d tx_hw_cons:%u cons:%u prod:%u curr:%u)", 479 "bnxt: ring busy w/ flush pending!\n"); 1959 * -EBUSY - completion ring does not have all the agg buffers yet 2231 /* In netpoll mode, if we are using a combined completion ring, we need to 2656 netdev_warn(bp->dev, "Ring monitor event, ring type %lu id 0x%x\n", 2663 netdev_warn(bp->dev, "Unknown RX agg ring id 0x%x\n", 2809 /* disable ring IRQ */ 2945 /* ACK completion ring before freeing tx ring and producing new 2947 * ring 3531 struct bnxt_ring_struct *ring; local 3591 struct bnxt_ring_struct *ring; local 3647 struct bnxt_ring_struct *ring; local 3692 struct bnxt_ring_struct *ring; local 3733 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local 3797 struct bnxt_ring_struct *ring; local 3828 struct bnxt_ring_struct *ring; local 3864 struct bnxt_ring_struct *ring; local 3940 struct bnxt_ring_struct *ring; local 3990 bnxt_init_rxbd_pages(struct bnxt_ring_struct *ring, u32 type) argument 4064 struct bnxt_ring_struct *ring; local 4105 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local 4153 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; local 5154 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local 5996 bnxt_cp_ring_from_grp(struct bnxt *bp, struct bnxt_ring_struct *ring) argument 6313 unsigned int ring = 0, grp_idx; local 6568 hwrm_ring_alloc_send_msg(struct bnxt *bp, struct bnxt_ring_struct *ring, u32 ring_type, u32 map_index) argument 6798 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local 6824 struct bnxt_ring_struct *ring; local 6853 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; local 6884 struct bnxt_ring_struct *ring = local 6904 hwrm_ring_free_send_msg(struct bnxt *bp, struct bnxt_ring_struct *ring, u32 ring_type, int cmpl_ring_id) argument 6947 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; local 6962 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; local 6984 struct bnxt_ring_struct *ring = &rxr->rx_agg_ring_struct; local 7012 struct bnxt_ring_struct *ring; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_migrate.c | 49 svm_migrate_gart_map(struct amdgpu_ring *ring, uint64_t npages, argument 52 struct amdgpu_device *adev = ring->adev; 82 amdgpu_ring_pad_ib(ring, &job->ibs[0]); 103 * @adev: amdgpu device the sdma ring running 113 * GART update and sdma uses same buf copy function ring, sdma is splited to 130 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local 143 r = svm_migrate_gart_map(ring, size, sys, &gart_d, 0); 146 r = svm_migrate_gart_map(ring, size, sys, &gart_s, 155 r = amdgpu_copy_buffer(ring, gart_s, gart_d, size * PAGE_SIZE,
|
/linux-master/drivers/usb/host/ |
H A D | xhci-trace.h | 111 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 112 TP_ARGS(ring, trb), 121 __entry->type = ring->type; 134 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 135 TP_ARGS(ring, trb) 139 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 140 TP_ARGS(ring, trb) 144 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 145 TP_ARGS(ring, trb) 149 TP_PROTO(struct xhci_ring *ring, struc [all...] |
H A D | xhci-ring.c | 15 * Consumer Cycle State (CCS), depending on ring function. 19 * 1. A ring is empty if enqueue == dequeue. This means there will always be at 20 * least one free TRB in the ring. This is useful if you want to turn that 21 * into a link TRB and expand the ring. 24 * link TRB had its toggle bit set, you may need to update the ring cycle 27 * 3. A ring is full if enqueue++ (for the definition of increment above) 32 * in a link TRB, it must toggle the ring cycle state. 34 * in a link TRB, it must toggle the ring cycle state. 37 * 1. Check if ring is full before you enqueue. 38 * 2. Write the ring cycl 98 last_trb_on_ring(struct xhci_ring *ring, struct xhci_segment *seg, union xhci_trb *trb) argument 116 unhandled_event_trb(struct xhci_ring *ring) argument 148 next_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_segment **seg, union xhci_trb **trb) argument 164 inc_deq(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 220 inc_enq(struct xhci_hcd *xhci, struct xhci_ring *ring, bool more_trbs_coming) argument 286 xhci_num_trbs_free(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 322 xhci_ring_expansion_needed(struct xhci_hcd *xhci, struct xhci_ring *ring, unsigned int num_trbs) argument 812 xhci_unmap_td_bounce_buffer(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_td *td) argument 898 struct xhci_ring *ring; local 999 struct xhci_ring *ring; local 1199 xhci_kill_ring_urbs(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 1224 struct xhci_ring *ring; local 2259 sum_trb_lengths(struct xhci_hcd *xhci, struct xhci_ring *ring, union xhci_trb *stop_trb) argument 3194 queue_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, bool more_trbs_coming, u32 field1, u32 field2, u32 field3, u32 field4) argument 3596 struct xhci_ring *ring; local [all...] |
/linux-master/net/sched/ |
H A D | sch_generic.c | 811 /* NULL ring is possible if destroy path is due to a failed 814 if (!q->ring.queue) 878 /* NULL ring is possible if destroy path is due to a failed 881 if (!q->ring.queue) 883 /* Destroy ring but no need to kfree_skb because a call to 886 ptr_ring_cleanup(&q->ring, NULL);
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_wed.c | 223 /* prefetch ring status */ 229 /* writeback ring status */ 345 /* prefetch ring status */ 351 /* writeback ring status */ 624 /* eagle E1 PCIE1 tx ring 22 flow control issue */ 933 mtk_wed_free_ring(struct mtk_wed_device *dev, struct mtk_wed_ring *ring) argument 935 if (!ring->desc) 938 dma_free_coherent(dev->hw->dev, ring->size * ring->desc_size, 939 ring 1008 mtk_wed_check_wfdma_rx_fill(struct mtk_wed_device *dev, struct mtk_wed_ring *ring) argument 1302 mtk_wed_rro_ring_alloc(struct mtk_wed_device *dev, struct mtk_wed_ring *ring, int size) argument 1356 } ring[2]; member in struct:__anon17 1542 mtk_wed_ring_reset(struct mtk_wed_ring *ring, int size, bool tx) argument 1860 mtk_wed_ring_alloc(struct mtk_wed_device *dev, struct mtk_wed_ring *ring, int size, u32 desc_size, bool tx) argument 2140 struct mtk_wed_ring *ring = &dev->rx_ring[i]; local 2213 struct mtk_wed_ring *ring = &dev->rx_rro_ring[i]; local 2224 struct mtk_wed_ring *ring = &dev->rx_page_ring[i]; local 2239 struct mtk_wed_ring *ring = &dev->rx_rro_ring[idx]; local 2252 struct mtk_wed_ring *ring = &dev->rx_page_ring[idx]; local 2265 struct mtk_wed_ring *ring = &dev->ind_cmd_ring; local 2462 struct mtk_wed_ring *ring = &dev->tx_ring[idx]; local 2524 struct mtk_wed_ring *ring = &dev->txfree_ring; local 2549 struct mtk_wed_ring *ring = &dev->rx_ring[idx]; local [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_tc_lib.c | 912 struct ice_rx_ring *ring = NULL; local 936 ring = ice_locate_rx_ring_using_queue(vsi, tc_fltr); 937 if (!ring) {
|