Searched refs:ring (Results 1 - 25 of 804) sorted by last modified time

1234567891011>>

/linux-master/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c345 /* RDMA/TDMA ring registers and accessors
409 unsigned int ring,
413 (DMA_RING_SIZE * ring) +
418 unsigned int ring, u32 val,
422 (DMA_RING_SIZE * ring) +
427 unsigned int ring,
431 (DMA_RING_SIZE * ring) +
436 unsigned int ring, u32 val,
440 (DMA_RING_SIZE * ring) +
672 * We'll map them to ring
408 bcmgenet_tdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument
417 bcmgenet_tdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument
426 bcmgenet_rdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument
435 bcmgenet_rdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument
818 struct bcmgenet_rx_ring *ring; local
840 bcmgenet_set_rx_coalesce(struct bcmgenet_rx_ring *ring, u32 usecs, u32 pkts) argument
855 bcmgenet_set_ring_rx_coalesce(struct bcmgenet_rx_ring *ring, struct ethtool_coalesce *ec) argument
1728 bcmgenet_get_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument
1745 bcmgenet_put_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument
1762 bcmgenet_rx_ring16_int_disable(struct bcmgenet_rx_ring *ring) argument
1768 bcmgenet_rx_ring16_int_enable(struct bcmgenet_rx_ring *ring) argument
1774 bcmgenet_rx_ring_int_disable(struct bcmgenet_rx_ring *ring) argument
1781 bcmgenet_rx_ring_int_enable(struct bcmgenet_rx_ring *ring) argument
1788 bcmgenet_tx_ring16_int_disable(struct bcmgenet_tx_ring *ring) argument
1794 bcmgenet_tx_ring16_int_enable(struct bcmgenet_tx_ring *ring) argument
1800 bcmgenet_tx_ring_int_enable(struct bcmgenet_tx_ring *ring) argument
1806 bcmgenet_tx_ring_int_disable(struct bcmgenet_tx_ring *ring) argument
1868 __bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument
1925 bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument
1939 struct bcmgenet_tx_ring *ring = local
2047 struct bcmgenet_tx_ring *ring = NULL; local
2229 bcmgenet_desc_rx(struct bcmgenet_rx_ring *ring, unsigned int budget) argument
2393 struct bcmgenet_rx_ring *ring = container_of(napi, local
2419 struct bcmgenet_rx_ring *ring = local
2429 bcmgenet_alloc_rx_buffers(struct bcmgenet_priv *priv, struct bcmgenet_rx_ring *ring) argument
2599 bcmgenet_init_dim(struct bcmgenet_rx_ring *ring, void (*cb)(struct work_struct *work)) argument
2611 bcmgenet_init_rx_coalesce(struct bcmgenet_rx_ring *ring) argument
2635 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; local
2694 struct bcmgenet_rx_ring *ring = &priv->rx_rings[index]; local
2750 struct bcmgenet_tx_ring *ring; local
2766 struct bcmgenet_tx_ring *ring; local
2780 struct bcmgenet_tx_ring *ring; local
2863 struct bcmgenet_rx_ring *ring; local
2879 struct bcmgenet_rx_ring *ring; local
2895 struct bcmgenet_rx_ring *ring; local
3498 bcmgenet_dump_tx_queue(struct bcmgenet_tx_ring *ring) argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_ttm.c170 * @ring: DMA ring to use for the copy
181 unsigned int window, struct amdgpu_ring *ring,
184 struct amdgpu_device *adev = ring->adev;
241 amdgpu_ring_pad_ib(ring, &job->ibs[0]);
294 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local
300 DRM_ERROR("Trying to move memory with ring turned off.\n");
317 0, ring, tmz, &cur_size, &from);
322 1, ring, tmz, &cur_size, &to);
326 r = amdgpu_copy_buffer(ring, fro
178 amdgpu_ttm_map_buffer(struct ttm_buffer_object *bo, struct ttm_resource *mem, struct amdgpu_res_cursor *mm_cur, unsigned int window, struct amdgpu_ring *ring, bool tmz, uint64_t *size, uint64_t *addr) argument
2066 struct amdgpu_ring *ring; local
2142 amdgpu_copy_buffer(struct amdgpu_ring *ring, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, struct dma_resv *resv, struct dma_fence **fence, bool direct_submit, bool vm_needs_flush, bool tmz) argument
2196 amdgpu_ttm_fill_mem(struct amdgpu_ring *ring, uint32_t src_data, uint64_t dst_addr, uint32_t byte_count, struct dma_resv *resv, struct dma_fence **fence, bool vm_needs_flush, bool delayed) argument
2240 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local
[all...]
H A Damdgpu_object.c754 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local
760 return amdgpu_copy_buffer(ring, shadow_addr, parent_addr,
H A Dvpe_v6_1.c67 base = vpe->ring.adev->reg_offset[VPE_HWIP][inst][0];
74 struct amdgpu_device *adev = vpe->ring.adev;
101 struct amdgpu_device *adev = vpe->ring.adev;
124 struct amdgpu_device *adev = vpe->ring.adev;
207 struct amdgpu_ring *ring = &vpe->ring; local
208 struct amdgpu_device *adev = ring->adev;
215 /* Set ring buffer size in dwords */
216 rb_bufsz = order_base_2(ring->ring_size / 4);
223 /* Initialize the ring buffe
[all...]
H A Dsdma_v5_2.c92 static unsigned sdma_v5_2_ring_init_cond_exec(struct amdgpu_ring *ring, argument
97 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE));
98 amdgpu_ring_write(ring, lower_32_bits(addr));
99 amdgpu_ring_write(ring, upper_32_bits(addr));
100 amdgpu_ring_write(ring, 1);
102 ret = ring->wptr & ring->buf_mask;
104 amdgpu_ring_write(ring, 0);
112 * @ring: amdgpu ring pointe
116 sdma_v5_2_ring_get_rptr(struct amdgpu_ring *ring) argument
134 sdma_v5_2_ring_get_wptr(struct amdgpu_ring *ring) argument
160 sdma_v5_2_ring_set_wptr(struct amdgpu_ring *ring) argument
194 sdma_v5_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
217 sdma_v5_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
252 sdma_v5_2_ring_emit_mem_sync(struct amdgpu_ring *ring) argument
277 sdma_v5_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
312 sdma_v5_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
475 struct amdgpu_ring *ring; local
809 sdma_v5_2_ring_test_ring(struct amdgpu_ring *ring) argument
885 sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1066 sdma_v5_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1091 sdma_v5_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1120 sdma_v5_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1126 sdma_v5_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1135 sdma_v5_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1149 sdma_v5_2_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
1213 struct amdgpu_ring *ring; local
1331 sdma_v5_2_ring_preempt_ib(struct amdgpu_ring *ring) argument
1638 sdma_v5_2_ring_begin_use(struct amdgpu_ring *ring) argument
1653 sdma_v5_2_ring_end_use(struct amdgpu_ring *ring) argument
[all...]
H A Dsdma_v4_4_2.c154 * @ring: amdgpu ring pointer
158 static uint64_t sdma_v4_4_2_ring_get_rptr(struct amdgpu_ring *ring) argument
163 rptr = READ_ONCE(*((u64 *)&ring->adev->wb.wb[ring->rptr_offs]));
172 * @ring: amdgpu ring pointer
176 static uint64_t sdma_v4_4_2_ring_get_wptr(struct amdgpu_ring *ring) argument
178 struct amdgpu_device *adev = ring->adev;
181 if (ring
203 sdma_v4_4_2_ring_set_wptr(struct amdgpu_ring *ring) argument
245 sdma_v4_4_2_page_ring_get_wptr(struct amdgpu_ring *ring) argument
269 sdma_v4_4_2_page_ring_set_wptr(struct amdgpu_ring *ring) argument
289 sdma_v4_4_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
312 sdma_v4_4_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
333 sdma_v4_4_2_wait_reg_mem(struct amdgpu_ring *ring, int mem_space, int hdp, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
365 sdma_v4_4_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
392 sdma_v4_4_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
596 sdma_v4_4_2_rb_cntl(struct amdgpu_ring *ring, uint32_t rb_cntl) argument
622 struct amdgpu_ring *ring = &adev->sdma.instance[i].ring; local
710 struct amdgpu_ring *ring = &adev->sdma.instance[i].page; local
867 struct amdgpu_ring *ring; local
954 sdma_v4_4_2_ring_test_ring(struct amdgpu_ring *ring) argument
1007 sdma_v4_4_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1159 sdma_v4_4_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1184 sdma_v4_4_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1207 sdma_v4_4_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1213 sdma_v4_4_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1222 sdma_v4_4_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1282 struct amdgpu_ring *ring; local
[all...]
H A Dgfx_v9_0.c758 static void gfx_v9_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume, bool usegds);
759 static u64 gfx_v9_0_ring_get_rptr_compute(struct amdgpu_ring *ring);
787 struct amdgpu_ring *ring)
789 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj);
790 uint64_t wptr_addr = ring->wptr_gpu_addr;
791 uint32_t eng_sel = ring->funcs->type == AMDGPU_RING_TYPE_GFX ? 4 : 0;
798 PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
799 PACKET3_MAP_QUEUES_PIPE(ring->pipe) |
800 PACKET3_MAP_QUEUES_ME((ring->me == 1 ? 0 : 1)) |
809 PACKET3_MAP_QUEUES_DOORBELL_OFFSET(ring
786 gfx_v9_0_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument
816 gfx_v9_0_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument
844 gfx_v9_0_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument
960 gfx_v9_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument
972 gfx_v9_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
994 gfx_v9_0_ring_test_ring(struct amdgpu_ring *ring) argument
1024 gfx_v9_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1969 struct amdgpu_ring *ring = &adev->gfx.compute_ring[ring_id]; local
2001 struct amdgpu_ring *ring; local
3025 struct amdgpu_ring *ring = &adev->gfx.gfx_ring[0]; local
3095 struct amdgpu_ring *ring; local
3217 gfx_v9_0_kiq_setting(struct amdgpu_ring *ring) argument
3231 gfx_v9_0_mqd_set_priority(struct amdgpu_ring *ring, struct v9_mqd *mqd) argument
3244 gfx_v9_0_mqd_init(struct amdgpu_ring *ring) argument
3376 gfx_v9_0_kiq_init_register(struct amdgpu_ring *ring) argument
3490 gfx_v9_0_kiq_fini_register(struct amdgpu_ring *ring) argument
3529 gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring) argument
3577 gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring) argument
3617 struct amdgpu_ring *ring; local
3641 struct amdgpu_ring *ring = NULL; local
3671 struct amdgpu_ring *ring; local
3926 struct amdgpu_ring *ring = &kiq->ring; local
4032 gfx_v9_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument
4299 struct amdgpu_ring *ring = &adev->gfx.compute_ring[0]; local
4346 struct amdgpu_ring *ring = &adev->gfx.compute_ring[0]; local
4904 gfx_v9_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned int vmid) argument
5073 gfx_v9_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument
5078 gfx_v9_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument
5094 gfx_v9_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument
5108 gfx_v9_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
5137 gfx_v9_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5178 gfx_v9_0_ring_patch_cntl(struct amdgpu_ring *ring, unsigned offset) argument
5187 gfx_v9_0_ring_patch_ce_meta(struct amdgpu_ring *ring, unsigned offset) argument
5219 gfx_v9_0_ring_patch_de_meta(struct amdgpu_ring *ring, unsigned offset) argument
5254 gfx_v9_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5289 gfx_v9_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
5330 gfx_v9_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
5341 gfx_v9_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
5354 gfx_v9_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument
5359 gfx_v9_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument
5371 gfx_v9_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument
5384 gfx_v9_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
5411 gfx_v9_ring_emit_sb(struct amdgpu_ring *ring) argument
5417 gfx_v9_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume) argument
5459 gfx_v9_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
5517 gfx_v9_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume, bool usegds) argument
5571 gfx_v9_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument
5580 gfx_v9_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument
5613 gfx_v9_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument
5628 gfx_v9_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument
5645 gfx_v9_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
5668 gfx_v9_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
5674 gfx_v9_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
5691 gfx_v9_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument
5893 struct amdgpu_ring *ring; local
5931 struct amdgpu_ring *ring; local
6765 gfx_v9_0_emit_mem_sync(struct amdgpu_ring *ring) argument
6784 gfx_v9_0_emit_wave_limit_cs(struct amdgpu_ring *ring, uint32_t pipe, bool enable) argument
6815 gfx_v9_0_emit_wave_limit(struct amdgpu_ring *ring, bool enable) argument
[all...]
H A Dgfx_v11_0.c133 static void gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume);
134 static void gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure);
135 static void gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg,
138 static void gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring,
161 struct amdgpu_ring *ring)
163 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj);
164 uint64_t wptr_addr = ring->wptr_gpu_addr;
167 switch (ring->funcs->type) {
189 PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
190 PACKET3_MAP_QUEUES_PIPE(ring
160 gfx11_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument
203 gfx11_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument
236 gfx11_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument
303 gfx_v11_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument
314 gfx_v11_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
336 gfx_v11_0_ring_test_ring(struct amdgpu_ring *ring) argument
377 gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
933 struct amdgpu_ring *ring; local
965 struct amdgpu_ring *ring; local
3151 struct amdgpu_ring *ring; local
3236 gfx_v11_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
3263 struct amdgpu_ring *ring; local
3589 gfx_v11_0_kiq_setting(struct amdgpu_ring *ring) argument
3704 gfx_v11_0_gfx_init_queue(struct amdgpu_ring *ring) argument
3735 struct amdgpu_ring *ring; local
3892 gfx_v11_0_kiq_init_register(struct amdgpu_ring *ring) argument
4002 gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring) argument
4041 gfx_v11_0_kcq_init_queue(struct amdgpu_ring *ring) argument
4072 struct amdgpu_ring *ring; local
4097 struct amdgpu_ring *ring = NULL; local
4128 struct amdgpu_ring *ring; local
4630 struct amdgpu_ring *ring; local
4687 gfx_v11_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument
5036 gfx_v11_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned vmid) argument
5211 gfx_v11_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument
5217 gfx_v11_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument
5233 gfx_v11_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument
5250 gfx_v11_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument
5256 gfx_v11_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument
5268 gfx_v11_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument
5282 gfx_v11_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
5311 gfx_v11_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5351 gfx_v11_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5390 gfx_v11_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
5428 gfx_v11_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
5438 gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument
5450 gfx_v11_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
5472 gfx_v11_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
5499 gfx_v11_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument
5519 gfx_v11_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument
5536 gfx_v11_0_ring_emit_gfx_shadow(struct amdgpu_ring *ring, u64 shadow_va, u64 csa_va, u64 gds_va, bool init_shadow, int vmid) argument
5627 gfx_v11_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
5674 gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument
5725 gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument
5734 gfx_v11_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument
5751 gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
5774 gfx_v11_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
5780 gfx_v11_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
5790 gfx_v11_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument
5942 struct amdgpu_ring *ring; local
6033 struct amdgpu_ring *ring; local
6136 gfx_v11_0_emit_mem_sync(struct amdgpu_ring *ring) argument
[all...]
H A Dgfx_v10_0.c52 * 1. Primary ring
53 * 2. Async ring
3489 static void gfx_v10_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume);
3490 static void gfx_v10_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume);
3491 static void gfx_v10_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure);
3495 static void gfx_v10_0_ring_invalidate_tlbs(struct amdgpu_ring *ring,
3517 struct amdgpu_ring *ring)
3519 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj);
3520 uint64_t wptr_addr = ring->wptr_gpu_addr;
3523 switch (ring
3516 gfx10_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument
3556 gfx10_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument
3589 gfx10_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument
3743 gfx_v10_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument
3754 gfx_v10_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
3776 gfx_v10_0_ring_test_ring(struct amdgpu_ring *ring) argument
3814 gfx_v10_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
4434 struct amdgpu_ring *ring; local
4465 struct amdgpu_ring *ring; local
5961 struct amdgpu_ring *ring; local
6050 gfx_v10_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
6097 struct amdgpu_ring *ring; local
6311 gfx_v10_0_kiq_setting(struct amdgpu_ring *ring) argument
6445 gfx_v10_0_gfx_init_queue(struct amdgpu_ring *ring) argument
6492 struct amdgpu_ring *ring; local
6635 gfx_v10_0_kiq_init_register(struct amdgpu_ring *ring) argument
6744 gfx_v10_0_kiq_init_queue(struct amdgpu_ring *ring) argument
6783 gfx_v10_0_kcq_init_queue(struct amdgpu_ring *ring) argument
6814 struct amdgpu_ring *ring; local
6838 struct amdgpu_ring *ring = NULL; local
6868 struct amdgpu_ring *ring; local
7406 gfx_v10_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument
7953 gfx_v10_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned int vmid) argument
8174 gfx_v10_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument
8180 gfx_v10_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument
8196 gfx_v10_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument
8241 gfx_v10_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument
8247 gfx_v10_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument
8259 gfx_v10_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument
8301 gfx_v10_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
8330 gfx_v10_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
8371 gfx_v10_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
8410 gfx_v10_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
8444 gfx_v10_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
8454 gfx_v10_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument
8466 gfx_v10_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument
8482 gfx_v10_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
8509 gfx_v10_0_ring_emit_sb(struct amdgpu_ring *ring) argument
8515 gfx_v10_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument
8549 gfx_v10_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument
8566 gfx_v10_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
8613 gfx_v10_0_ring_emit_ce_meta(struct amdgpu_ring *ring, bool resume) argument
8653 gfx_v10_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument
8704 gfx_v10_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument
8713 gfx_v10_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument
8730 gfx_v10_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
8753 gfx_v10_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
8759 gfx_v10_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
8777 gfx_v10_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned int vmid) argument
8933 struct amdgpu_ring *ring; local
9024 struct amdgpu_ring *ring; local
9078 struct amdgpu_ring *ring = &(adev->gfx.kiq[0].ring); local
9122 struct amdgpu_ring *ring = &(adev->gfx.kiq[0].ring); local
9134 gfx_v10_0_emit_mem_sync(struct amdgpu_ring *ring) argument
[all...]
H A Damdgpu_vpe.c123 struct amdgpu_device *adev = vpe->ring.adev;
224 struct amdgpu_device *adev = vpe->ring.adev;
267 struct amdgpu_ring *ring = &vpe->ring; local
270 ring->ring_obj = NULL;
271 ring->use_doorbell = true;
272 ring->vm_hub = AMDGPU_MMHUB0(0);
273 ring->doorbell_index = (adev->doorbell_index.vpe_ring << 1);
274 snprintf(ring->name, 4, "vpe");
276 ret = amdgpu_ring_init(adev, ring, 102
445 vpe_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
457 vpe_get_csa_mc_addr(struct amdgpu_ring *ring, uint32_t vmid) argument
472 vpe_ring_emit_pred_exec(struct amdgpu_ring *ring, uint32_t device_select, uint32_t exec_count) argument
484 vpe_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
503 vpe_ring_emit_fence(struct amdgpu_ring *ring, uint64_t addr, uint64_t seq, unsigned int flags) argument
527 vpe_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
547 vpe_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
556 vpe_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
573 vpe_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument
579 vpe_ring_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument
594 vpe_ring_preempt_ib(struct amdgpu_ring *ring) argument
662 vpe_ring_get_rptr(struct amdgpu_ring *ring) argument
681 vpe_ring_get_wptr(struct amdgpu_ring *ring) argument
700 vpe_ring_set_wptr(struct amdgpu_ring *ring) argument
734 vpe_ring_test_ring(struct amdgpu_ring *ring) argument
776 vpe_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
830 vpe_ring_begin_use(struct amdgpu_ring *ring) argument
855 vpe_ring_end_use(struct amdgpu_ring *ring) argument
[all...]
H A Damdgpu_umsch_mm.c269 uint32_t *ring = test->ring_data_cpu_addr + local
285 ring[0] = VPE_CMD_HEADER(VPE_CMD_OPCODE_INDIRECT, 0);
286 ring[1] = (ib_gpu_addr & 0xffffffe0);
287 ring[2] = upper_32_bits(ib_gpu_addr);
288 ring[3] = 4;
289 ring[4] = 0;
290 ring[5] = 0;
491 struct amdgpu_ring *ring = &umsch->ring; local
493 if (amdgpu_ring_alloc(ring, ndw
504 struct amdgpu_ring *ring = &umsch->ring; local
518 umsch_mm_ring_set_wptr(struct amdgpu_ring *ring) argument
529 umsch_mm_ring_get_rptr(struct amdgpu_ring *ring) argument
537 umsch_mm_ring_get_wptr(struct amdgpu_ring *ring) argument
559 struct amdgpu_ring *ring = &umsch->ring; local
[all...]
H A Damdgpu_mes.c171 "(%d) ring trail_fence_offs wb alloc failed\n", r);
596 amdgpu_gfx_select_me_pipe_q(adev, p->ring->me, p->ring->pipe, 0, 0, 0);
704 queue->ring = qprops->ring;
710 queue->ring->mqd_ptr = queue->mqd_cpu_ptr;
782 struct amdgpu_ring *ring,
790 queue_input.queue_type = ring->funcs->type;
791 queue_input.doorbell_offset = ring->doorbell_index;
792 queue_input.pipe_id = ring
781 amdgpu_mes_unmap_legacy_queue(struct amdgpu_device *adev, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument
973 amdgpu_mes_ring_to_queue_props(struct amdgpu_device *adev, struct amdgpu_ring *ring, struct amdgpu_mes_queue_properties *props) argument
1007 amdgpu_mes_ctx_get_offs(struct amdgpu_ring *ring, unsigned int id_offs) argument
1032 struct amdgpu_ring *ring; local
1128 amdgpu_mes_remove_ring(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
1310 struct amdgpu_ring *ring; local
1345 struct amdgpu_ring *ring; local
[all...]
/linux-master/drivers/net/ethernet/wangxun/libwx/
H A Dwx_lib.c228 /* the page has been released from the ring */
326 * @rx_ring: ring to place buffers on
375 /* update next to alloc since we have filled the ring */
388 u16 wx_desc_unused(struct wx_ring *ring) argument
390 u16 ntc = ring->next_to_clean;
391 u16 ntu = ring->next_to_use;
393 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1;
398 * @rx_ring: Rx ring being processed
458 * @rx_ring: rx descriptor ring packet is being transacted on
499 static void wx_rx_hash(struct wx_ring *ring, argument
525 wx_rx_checksum(struct wx_ring *ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument
571 wx_rx_vlan(struct wx_ring *ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument
694 wx_txring_txq(const struct wx_ring *ring) argument
837 struct wx_ring *ring; local
1711 wx_add_ring(struct wx_ring *ring, struct wx_ring_container *head) argument
1737 struct wx_ring *ring; local
1843 struct wx_ring *ring; local
2158 struct wx_ring *ring; local
2637 struct wx_ring *ring = READ_ONCE(wx->rx_ring[i]); local
2653 struct wx_ring *ring = READ_ONCE(wx->tx_ring[i]); local
[all...]
/linux-master/drivers/net/ethernet/renesas/
H A Dravb_main.c330 /* Build RX ring buffer */
370 /* Build TX ring buffer */
856 /* Refill the RX ring buffers. */
967 /* Refill the RX ring buffers. */
1624 struct ethtool_ringparam *ring,
1630 ring->rx_max_pending = BE_RX_RING_MAX;
1631 ring->tx_max_pending = BE_TX_RING_MAX;
1632 ring->rx_pending = priv->num_rx_ring[RAVB_BE];
1633 ring->tx_pending = priv->num_tx_ring[RAVB_BE];
1637 struct ethtool_ringparam *ring,
1623 ravb_get_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument
1636 ravb_set_ringparam(struct net_device *ndev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument
[all...]
/linux-master/drivers/net/ethernet/intel/igc/
H A Digc_main.c192 * @tx_ring: ring to be cleaned
257 /* Zero out the buffer ring */
261 /* Zero out the descriptor ring */
271 * @tx_ring: Tx descriptor ring for a specific queue
319 static void igc_disable_tx_ring_hw(struct igc_ring *ring) argument
321 struct igc_hw *hw = &ring->q_vector->adapter->hw;
322 u8 idx = ring->reg_idx;
348 * @tx_ring: tx descriptor ring (for a specific queue) to setup
380 netdev_err(ndev, "Unable to allocate memory for Tx descriptor ring\n");
415 /* Free all the Rx ring sk_buff
443 igc_clean_rx_ring_xsk_pool(struct igc_ring *ring) argument
462 igc_clean_rx_ring(struct igc_ring *ring) argument
608 igc_get_xsk_pool(struct igc_adapter *adapter, struct igc_ring *ring) argument
625 igc_configure_rx_ring(struct igc_adapter *adapter, struct igc_ring *ring) argument
728 igc_configure_tx_ring(struct igc_adapter *adapter, struct igc_ring *ring) argument
1038 igc_tx_launchtime(struct igc_ring *ring, ktime_t txtime, bool *first_flag, bool *insert_empty) argument
1084 igc_init_empty_frame(struct igc_ring *ring, struct igc_tx_buffer *buffer, struct sk_buff *skb) argument
1110 igc_init_tx_empty_descriptor(struct igc_ring *ring, struct sk_buff *skb, struct igc_tx_buffer *first) argument
1715 igc_rx_checksum(struct igc_ring *ring, union igc_adv_rx_desc *rx_desc, struct sk_buff *skb) argument
1775 igc_rx_hash(struct igc_ring *ring, union igc_adv_rx_desc *rx_desc, struct sk_buff *skb) argument
1892 igc_get_rx_frame_truesize(struct igc_ring *ring, unsigned int size) argument
2292 igc_alloc_rx_buffers_zc(struct igc_ring *ring, u16 count) argument
2352 igc_xdp_init_tx_descriptor(struct igc_ring *ring, struct xdp_frame *xdpf) argument
2468 struct igc_ring *ring; local
2535 igc_flush_tx_descriptors(struct igc_ring *ring) argument
2549 struct igc_ring *ring; local
2567 struct igc_ring *ring = q_vector->rx.ring; local
2706 igc_construct_skb_zc(struct igc_ring *ring, struct xdp_buff *xdp) argument
2736 struct igc_ring *ring = q_vector->rx.ring; local
2767 struct igc_ring *ring = q_vector->rx.ring; local
2866 struct igc_ring *ring = q_vector->tx.ring; local
2877 igc_xdp_xmit_zc(struct igc_ring *ring) argument
3944 struct igc_ring *ring = adapter->rx_ring[i]; local
4525 igc_add_ring(struct igc_ring *ring, struct igc_ring_container *head) argument
4612 struct igc_ring *ring; local
4909 struct igc_ring *ring = adapter->rx_ring[i]; local
4935 struct igc_ring *ring = adapter->tx_ring[i]; local
6048 struct igc_ring *ring; local
6149 struct igc_ring *ring = adapter->tx_ring[i]; local
6161 struct igc_ring *ring = adapter->tx_ring[i]; local
6261 struct igc_ring *ring = adapter->tx_ring[i]; local
6289 struct igc_ring *ring = adapter->tx_ring[i]; local
6312 struct igc_ring *ring = adapter->tx_ring[i]; local
6346 struct igc_ring *ring; local
6476 struct igc_ring *ring; local
6526 struct igc_ring *ring; local
7393 igc_disable_rx_ring_hw(struct igc_ring *ring) argument
7405 igc_disable_rx_ring(struct igc_ring *ring) argument
7411 igc_enable_rx_ring(struct igc_ring *ring) argument
7423 igc_disable_tx_ring(struct igc_ring *ring) argument
7429 igc_enable_tx_ring(struct igc_ring *ring) argument
[all...]
H A Digc.h100 struct igc_ring *ring; /* pointer to linked list of rings */ member in struct:igc_ring_container
105 u8 itr; /* current ITR setting for ring */
116 void *desc; /* descriptor ring memory */
117 unsigned long flags; /* ring specific flags */
118 void __iomem *tail; /* pointer to ring tail register */
119 dma_addr_t dma; /* phys address of the ring */
120 unsigned int size; /* length of desc. ring in bytes */
122 u16 count; /* number of desc. in the ring */
123 u8 queue_index; /* logical index of the ring*/
124 u8 reg_idx; /* physical index of the ring */
630 igc_desc_unused(const struct igc_ring *ring) argument
680 igc_rx_bufsz(struct igc_ring *ring) argument
692 igc_rx_pg_order(struct igc_ring *ring) argument
[all...]
/linux-master/drivers/net/ethernet/intel/iavf/
H A Diavf_main.c433 if (!q_vector->tx.ring && !q_vector->rx.ring)
455 rx_ring->next = q_vector->rx.ring;
457 q_vector->rx.ring = rx_ring;
481 tx_ring->next = q_vector->tx.ring;
483 q_vector->tx.ring = tx_ring;
499 * one vector per ring/queue, but on a constrained vector budget, we
577 if (q_vector->tx.ring && q_vector->rx.ring) {
581 } else if (q_vector->rx.ring) {
1246 struct iavf_ring *ring = &adapter->rx_rings[i]; local
[all...]
/linux-master/drivers/net/ethernet/intel/i40e/
H A Di40e_main.c432 * i40e_get_netdev_stats_struct_tx - populate stats from a Tx ring
433 * @ring: Tx ring to get statistics from
436 static void i40e_get_netdev_stats_struct_tx(struct i40e_ring *ring, argument
443 start = u64_stats_fetch_begin(&ring->syncp);
444 packets = ring->stats.packets;
445 bytes = ring->stats.bytes;
446 } while (u64_stats_fetch_retry(&ring->syncp, start));
466 struct i40e_ring *ring; local
480 ring
3415 i40e_config_xps_tx_ring(struct i40e_ring *ring) argument
3437 i40e_xsk_pool(struct i40e_ring *ring) argument
3457 i40e_configure_tx_ring(struct i40e_ring *ring) argument
3581 i40e_configure_rx_ring(struct i40e_ring *ring) argument
5109 struct i40e_ring *ring; local
11639 struct i40e_ring *ring; local
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt.c398 netdev_err(bp->dev, "Invalid Tx completion (ring:%d tx_hw_cons:%u cons:%u prod:%u curr:%u)",
479 "bnxt: ring busy w/ flush pending!\n");
1959 * -EBUSY - completion ring does not have all the agg buffers yet
2231 /* In netpoll mode, if we are using a combined completion ring, we need to
2656 netdev_warn(bp->dev, "Ring monitor event, ring type %lu id 0x%x\n",
2663 netdev_warn(bp->dev, "Unknown RX agg ring id 0x%x\n",
2809 /* disable ring IRQ */
2945 /* ACK completion ring before freeing tx ring and producing new
2947 * ring
3531 struct bnxt_ring_struct *ring; local
3591 struct bnxt_ring_struct *ring; local
3647 struct bnxt_ring_struct *ring; local
3692 struct bnxt_ring_struct *ring; local
3733 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local
3797 struct bnxt_ring_struct *ring; local
3828 struct bnxt_ring_struct *ring; local
3864 struct bnxt_ring_struct *ring; local
3940 struct bnxt_ring_struct *ring; local
3990 bnxt_init_rxbd_pages(struct bnxt_ring_struct *ring, u32 type) argument
4064 struct bnxt_ring_struct *ring; local
4105 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local
4153 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; local
5154 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local
5996 bnxt_cp_ring_from_grp(struct bnxt *bp, struct bnxt_ring_struct *ring) argument
6313 unsigned int ring = 0, grp_idx; local
6568 hwrm_ring_alloc_send_msg(struct bnxt *bp, struct bnxt_ring_struct *ring, u32 ring_type, u32 map_index) argument
6798 struct bnxt_ring_struct *ring = &cpr->cp_ring_struct; local
6824 struct bnxt_ring_struct *ring; local
6853 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; local
6884 struct bnxt_ring_struct *ring = local
6904 hwrm_ring_free_send_msg(struct bnxt *bp, struct bnxt_ring_struct *ring, u32 ring_type, int cmpl_ring_id) argument
6947 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; local
6962 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; local
6984 struct bnxt_ring_struct *ring = &rxr->rx_agg_ring_struct; local
7012 struct bnxt_ring_struct *ring; local
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_migrate.c49 svm_migrate_gart_map(struct amdgpu_ring *ring, uint64_t npages, argument
52 struct amdgpu_device *adev = ring->adev;
82 amdgpu_ring_pad_ib(ring, &job->ibs[0]);
103 * @adev: amdgpu device the sdma ring running
113 * GART update and sdma uses same buf copy function ring, sdma is splited to
130 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; local
143 r = svm_migrate_gart_map(ring, size, sys, &gart_d, 0);
146 r = svm_migrate_gart_map(ring, size, sys, &gart_s,
155 r = amdgpu_copy_buffer(ring, gart_s, gart_d, size * PAGE_SIZE,
/linux-master/drivers/usb/host/
H A Dxhci-trace.h111 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
112 TP_ARGS(ring, trb),
121 __entry->type = ring->type;
134 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
135 TP_ARGS(ring, trb)
139 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
140 TP_ARGS(ring, trb)
144 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
145 TP_ARGS(ring, trb)
149 TP_PROTO(struct xhci_ring *ring, struc
[all...]
H A Dxhci-ring.c15 * Consumer Cycle State (CCS), depending on ring function.
19 * 1. A ring is empty if enqueue == dequeue. This means there will always be at
20 * least one free TRB in the ring. This is useful if you want to turn that
21 * into a link TRB and expand the ring.
24 * link TRB had its toggle bit set, you may need to update the ring cycle
27 * 3. A ring is full if enqueue++ (for the definition of increment above)
32 * in a link TRB, it must toggle the ring cycle state.
34 * in a link TRB, it must toggle the ring cycle state.
37 * 1. Check if ring is full before you enqueue.
38 * 2. Write the ring cycl
98 last_trb_on_ring(struct xhci_ring *ring, struct xhci_segment *seg, union xhci_trb *trb) argument
116 unhandled_event_trb(struct xhci_ring *ring) argument
148 next_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_segment **seg, union xhci_trb **trb) argument
164 inc_deq(struct xhci_hcd *xhci, struct xhci_ring *ring) argument
220 inc_enq(struct xhci_hcd *xhci, struct xhci_ring *ring, bool more_trbs_coming) argument
286 xhci_num_trbs_free(struct xhci_hcd *xhci, struct xhci_ring *ring) argument
322 xhci_ring_expansion_needed(struct xhci_hcd *xhci, struct xhci_ring *ring, unsigned int num_trbs) argument
812 xhci_unmap_td_bounce_buffer(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_td *td) argument
898 struct xhci_ring *ring; local
999 struct xhci_ring *ring; local
1199 xhci_kill_ring_urbs(struct xhci_hcd *xhci, struct xhci_ring *ring) argument
1224 struct xhci_ring *ring; local
2259 sum_trb_lengths(struct xhci_hcd *xhci, struct xhci_ring *ring, union xhci_trb *stop_trb) argument
3194 queue_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, bool more_trbs_coming, u32 field1, u32 field2, u32 field3, u32 field4) argument
3596 struct xhci_ring *ring; local
[all...]
/linux-master/net/sched/
H A Dsch_generic.c811 /* NULL ring is possible if destroy path is due to a failed
814 if (!q->ring.queue)
878 /* NULL ring is possible if destroy path is due to a failed
881 if (!q->ring.queue)
883 /* Destroy ring but no need to kfree_skb because a call to
886 ptr_ring_cleanup(&q->ring, NULL);
/linux-master/drivers/net/ethernet/mediatek/
H A Dmtk_wed.c223 /* prefetch ring status */
229 /* writeback ring status */
345 /* prefetch ring status */
351 /* writeback ring status */
624 /* eagle E1 PCIE1 tx ring 22 flow control issue */
933 mtk_wed_free_ring(struct mtk_wed_device *dev, struct mtk_wed_ring *ring) argument
935 if (!ring->desc)
938 dma_free_coherent(dev->hw->dev, ring->size * ring->desc_size,
939 ring
1008 mtk_wed_check_wfdma_rx_fill(struct mtk_wed_device *dev, struct mtk_wed_ring *ring) argument
1302 mtk_wed_rro_ring_alloc(struct mtk_wed_device *dev, struct mtk_wed_ring *ring, int size) argument
1356 } ring[2]; member in struct:__anon17
1542 mtk_wed_ring_reset(struct mtk_wed_ring *ring, int size, bool tx) argument
1860 mtk_wed_ring_alloc(struct mtk_wed_device *dev, struct mtk_wed_ring *ring, int size, u32 desc_size, bool tx) argument
2140 struct mtk_wed_ring *ring = &dev->rx_ring[i]; local
2213 struct mtk_wed_ring *ring = &dev->rx_rro_ring[i]; local
2224 struct mtk_wed_ring *ring = &dev->rx_page_ring[i]; local
2239 struct mtk_wed_ring *ring = &dev->rx_rro_ring[idx]; local
2252 struct mtk_wed_ring *ring = &dev->rx_page_ring[idx]; local
2265 struct mtk_wed_ring *ring = &dev->ind_cmd_ring; local
2462 struct mtk_wed_ring *ring = &dev->tx_ring[idx]; local
2524 struct mtk_wed_ring *ring = &dev->txfree_ring; local
2549 struct mtk_wed_ring *ring = &dev->rx_ring[idx]; local
[all...]
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_tc_lib.c912 struct ice_rx_ring *ring = NULL; local
936 ring = ice_locate_rx_ring_using_queue(vsi, tc_fltr);
937 if (!ring) {

Completed in 607 milliseconds

1234567891011>>