Searched refs:ring (Results 151 - 175 of 804) sorted by relevance

1234567891011>>

/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Djpeg_v3_0.c47 * Set ring and irq function pointers
85 struct amdgpu_ring *ring; local
102 ring = adev->jpeg.inst->ring_dec;
103 ring->use_doorbell = true;
104 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1;
105 ring->vm_hub = AMDGPU_MMHUB0(0);
106 sprintf(ring->name, "jpeg_dec");
107 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst->irq, 0,
148 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local
151 adev->nbio.funcs->vcn_doorbell_range(adev, ring
334 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local
414 jpeg_v3_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
428 jpeg_v3_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
445 jpeg_v3_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
[all...]
H A Dsdma_v4_0.c602 * @ring: amdgpu ring pointer
606 static uint64_t sdma_v4_0_ring_get_rptr(struct amdgpu_ring *ring) argument
611 rptr = ((u64 *)ring->rptr_cpu_addr);
620 * @ring: amdgpu ring pointer
624 static uint64_t sdma_v4_0_ring_get_wptr(struct amdgpu_ring *ring) argument
626 struct amdgpu_device *adev = ring->adev;
629 if (ring->use_doorbell) {
631 wptr = READ_ONCE(*((u64 *)ring
651 sdma_v4_0_ring_set_wptr(struct amdgpu_ring *ring) argument
693 sdma_v4_0_page_ring_get_wptr(struct amdgpu_ring *ring) argument
717 sdma_v4_0_page_ring_set_wptr(struct amdgpu_ring *ring) argument
737 sdma_v4_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
760 sdma_v4_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
781 sdma_v4_0_wait_reg_mem(struct amdgpu_ring *ring, int mem_space, int hdp, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
813 sdma_v4_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
839 sdma_v4_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
1019 sdma_v4_0_rb_cntl(struct amdgpu_ring *ring, uint32_t rb_cntl) argument
1044 struct amdgpu_ring *ring = &adev->sdma.instance[i].ring; local
1129 struct amdgpu_ring *ring = &adev->sdma.instance[i].page; local
1337 struct amdgpu_ring *ring; local
1416 sdma_v4_0_ring_test_ring(struct amdgpu_ring *ring) argument
1469 sdma_v4_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1621 sdma_v4_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1646 sdma_v4_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1669 sdma_v4_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1675 sdma_v4_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1684 sdma_v4_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1750 struct amdgpu_ring *ring; local
[all...]
H A Dgfx_v11_0.c133 static void gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume);
134 static void gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure);
135 static void gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg,
138 static void gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring,
161 struct amdgpu_ring *ring)
163 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj);
164 uint64_t wptr_addr = ring->wptr_gpu_addr;
167 switch (ring->funcs->type) {
189 PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
190 PACKET3_MAP_QUEUES_PIPE(ring
160 gfx11_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument
203 gfx11_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument
236 gfx11_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument
303 gfx_v11_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument
314 gfx_v11_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument
336 gfx_v11_0_ring_test_ring(struct amdgpu_ring *ring) argument
377 gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
933 struct amdgpu_ring *ring; local
965 struct amdgpu_ring *ring; local
3151 struct amdgpu_ring *ring; local
3236 gfx_v11_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
3263 struct amdgpu_ring *ring; local
3589 gfx_v11_0_kiq_setting(struct amdgpu_ring *ring) argument
3704 gfx_v11_0_gfx_init_queue(struct amdgpu_ring *ring) argument
3735 struct amdgpu_ring *ring; local
3892 gfx_v11_0_kiq_init_register(struct amdgpu_ring *ring) argument
4002 gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring) argument
4041 gfx_v11_0_kcq_init_queue(struct amdgpu_ring *ring) argument
4072 struct amdgpu_ring *ring; local
4097 struct amdgpu_ring *ring = NULL; local
4128 struct amdgpu_ring *ring; local
4630 struct amdgpu_ring *ring; local
4687 gfx_v11_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument
5036 gfx_v11_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned vmid) argument
5211 gfx_v11_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument
5217 gfx_v11_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument
5233 gfx_v11_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument
5250 gfx_v11_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument
5256 gfx_v11_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument
5268 gfx_v11_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument
5282 gfx_v11_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
5311 gfx_v11_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5351 gfx_v11_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
5390 gfx_v11_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
5428 gfx_v11_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
5438 gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument
5450 gfx_v11_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
5472 gfx_v11_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
5499 gfx_v11_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument
5519 gfx_v11_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument
5536 gfx_v11_0_ring_emit_gfx_shadow(struct amdgpu_ring *ring, u64 shadow_va, u64 csa_va, u64 gds_va, bool init_shadow, int vmid) argument
5627 gfx_v11_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
5674 gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument
5725 gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument
5734 gfx_v11_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument
5751 gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
5774 gfx_v11_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
5780 gfx_v11_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
5790 gfx_v11_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument
5942 struct amdgpu_ring *ring; local
6033 struct amdgpu_ring *ring; local
6136 gfx_v11_0_emit_mem_sync(struct amdgpu_ring *ring) argument
[all...]
H A Djpeg_v4_0_5.c54 static void jpeg_v4_0_5_dec_ring_set_wptr(struct amdgpu_ring *ring);
66 * Set ring and irq function pointers
104 struct amdgpu_ring *ring; local
142 ring = adev->jpeg.inst[i].ring_dec;
143 ring->use_doorbell = true;
144 ring->vm_hub = AMDGPU_MMHUB0(0);
145 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1 + 8 * i;
146 sprintf(ring->name, "jpeg_dec_%d", i);
147 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst[i].irq,
189 struct amdgpu_ring *ring; local
399 struct amdgpu_ring *ring = adev->jpeg.inst[inst_idx].ring_dec; local
479 struct amdgpu_ring *ring; local
585 jpeg_v4_0_5_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
599 jpeg_v4_0_5_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
616 jpeg_v4_0_5_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
[all...]
H A Dpsp_v10_0.c75 struct psp_ring *ring = &psp->km_ring; local
78 /* Write low address of the ring to C2PMSG_69 */
79 psp_ring_reg = lower_32_bits(ring->ring_mem_mc_addr);
81 /* Write high address of the ring to C2PMSG_70 */
82 psp_ring_reg = upper_32_bits(ring->ring_mem_mc_addr);
84 /* Write size of ring to C2PMSG_71 */
85 psp_ring_reg = ring->ring_size;
87 /* Write the ring initialization command to C2PMSG_64 */
109 /* Write the ring destroy command to C2PMSG_64 */
127 struct psp_ring *ring local
[all...]
H A Djpeg_v4_0.c46 static void jpeg_v4_0_dec_ring_set_wptr(struct amdgpu_ring *ring);
53 * Set ring and irq function pointers
80 struct amdgpu_ring *ring; local
109 ring = adev->jpeg.inst->ring_dec;
110 ring->use_doorbell = true;
111 ring->doorbell_index = amdgpu_sriov_vf(adev) ? (((adev->doorbell_index.vcn.vcn_ring0_1) << 1) + 4) : ((adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1);
112 ring->vm_hub = AMDGPU_MMHUB0(0);
114 sprintf(ring->name, "jpeg_dec");
115 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst->irq, 0,
160 struct amdgpu_ring *ring local
368 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local
413 struct amdgpu_ring *ring; local
571 jpeg_v4_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
585 jpeg_v4_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
602 jpeg_v4_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
[all...]
H A Damdgpu_ids.c190 * @ring: ring we want to submit job to
197 static int amdgpu_vmid_grab_idle(struct amdgpu_ring *ring, argument
201 struct amdgpu_device *adev = ring->adev;
202 unsigned vmhub = ring->vm_hub;
207 if (!dma_fence_is_signaled(ring->vmid_wait)) {
208 *fence = dma_fence_get(ring->vmid_wait);
221 NULL : ring;
231 u64 fence_context = adev->vm_manager.fence_context + ring->idx;
232 unsigned seqno = ++adev->vm_manager.seqno[ring
270 amdgpu_vmid_grab_reserved(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_vmid **id, struct dma_fence **fence) argument
331 amdgpu_vmid_grab_used(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_vmid **id, struct dma_fence **fence) argument
394 amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct dma_fence **fence) argument
[all...]
H A Damdgpu_vcn.c280 static bool amdgpu_vcn_using_unified_queue(struct amdgpu_ring *ring) argument
282 struct amdgpu_device *adev = ring->adev;
429 void amdgpu_vcn_ring_begin_use(struct amdgpu_ring *ring) argument
431 struct amdgpu_device *adev = ring->adev;
450 if (ring->funcs->type == AMDGPU_RING_TYPE_VCN_ENC) {
451 atomic_inc(&adev->vcn.inst[ring->me].dpg_enc_submission_cnt);
458 fences += amdgpu_fence_count_emitted(&adev->vcn.inst[ring->me].ring_enc[i]);
460 if (fences || atomic_read(&adev->vcn.inst[ring->me].dpg_enc_submission_cnt))
466 adev->vcn.pause_dpg_mode(adev, ring->me, &new_state);
471 void amdgpu_vcn_ring_end_use(struct amdgpu_ring *ring) argument
482 amdgpu_vcn_dec_ring_test_ring(struct amdgpu_ring *ring) argument
513 amdgpu_vcn_dec_sw_ring_test_ring(struct amdgpu_ring *ring) argument
544 amdgpu_vcn_dec_send_msg(struct amdgpu_ring *ring, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument
593 amdgpu_vcn_dec_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) argument
628 amdgpu_vcn_dec_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) argument
655 amdgpu_vcn_dec_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
717 amdgpu_vcn_dec_sw_send_msg(struct amdgpu_ring *ring, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument
787 amdgpu_vcn_dec_sw_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
819 amdgpu_vcn_enc_ring_test_ring(struct amdgpu_ring *ring) argument
850 amdgpu_vcn_enc_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument
917 amdgpu_vcn_enc_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument
984 amdgpu_vcn_enc_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1019 amdgpu_vcn_unified_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1036 amdgpu_vcn_get_enc_ring_prio(int ring) argument
[all...]
/linux-master/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c345 /* RDMA/TDMA ring registers and accessors
409 unsigned int ring,
413 (DMA_RING_SIZE * ring) +
418 unsigned int ring, u32 val,
422 (DMA_RING_SIZE * ring) +
427 unsigned int ring,
431 (DMA_RING_SIZE * ring) +
436 unsigned int ring, u32 val,
440 (DMA_RING_SIZE * ring) +
672 * We'll map them to ring
408 bcmgenet_tdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument
417 bcmgenet_tdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument
426 bcmgenet_rdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument
435 bcmgenet_rdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument
818 struct bcmgenet_rx_ring *ring; local
840 bcmgenet_set_rx_coalesce(struct bcmgenet_rx_ring *ring, u32 usecs, u32 pkts) argument
855 bcmgenet_set_ring_rx_coalesce(struct bcmgenet_rx_ring *ring, struct ethtool_coalesce *ec) argument
1728 bcmgenet_get_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument
1745 bcmgenet_put_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument
1762 bcmgenet_rx_ring16_int_disable(struct bcmgenet_rx_ring *ring) argument
1768 bcmgenet_rx_ring16_int_enable(struct bcmgenet_rx_ring *ring) argument
1774 bcmgenet_rx_ring_int_disable(struct bcmgenet_rx_ring *ring) argument
1781 bcmgenet_rx_ring_int_enable(struct bcmgenet_rx_ring *ring) argument
1788 bcmgenet_tx_ring16_int_disable(struct bcmgenet_tx_ring *ring) argument
1794 bcmgenet_tx_ring16_int_enable(struct bcmgenet_tx_ring *ring) argument
1800 bcmgenet_tx_ring_int_enable(struct bcmgenet_tx_ring *ring) argument
1806 bcmgenet_tx_ring_int_disable(struct bcmgenet_tx_ring *ring) argument
1868 __bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument
1925 bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument
1939 struct bcmgenet_tx_ring *ring = local
2047 struct bcmgenet_tx_ring *ring = NULL; local
2229 bcmgenet_desc_rx(struct bcmgenet_rx_ring *ring, unsigned int budget) argument
2393 struct bcmgenet_rx_ring *ring = container_of(napi, local
2419 struct bcmgenet_rx_ring *ring = local
2429 bcmgenet_alloc_rx_buffers(struct bcmgenet_priv *priv, struct bcmgenet_rx_ring *ring) argument
2599 bcmgenet_init_dim(struct bcmgenet_rx_ring *ring, void (*cb)(struct work_struct *work)) argument
2611 bcmgenet_init_rx_coalesce(struct bcmgenet_rx_ring *ring) argument
2635 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; local
2694 struct bcmgenet_rx_ring *ring = &priv->rx_rings[index]; local
2750 struct bcmgenet_tx_ring *ring; local
2766 struct bcmgenet_tx_ring *ring; local
2780 struct bcmgenet_tx_ring *ring; local
2863 struct bcmgenet_rx_ring *ring; local
2879 struct bcmgenet_rx_ring *ring; local
2895 struct bcmgenet_rx_ring *ring; local
3498 bcmgenet_dump_tx_queue(struct bcmgenet_tx_ring *ring) argument
[all...]
/linux-master/drivers/usb/cdns3/
H A Dcdnsp-mem.c24 * Allocates a generic ring segment from the ring pool, sets the dma address,
133 * Link the ring to the new segments.
134 * Set Toggle Cycle for the new ring if needed.
137 struct cdnsp_ring *ring,
144 if (!ring || !first || !last)
147 next = ring->enq_seg->next;
148 cdnsp_link_segments(pdev, ring->enq_seg, first, ring->type);
149 cdnsp_link_segments(pdev, last, next, ring
136 cdnsp_link_rings(struct cdnsp_device *pdev, struct cdnsp_ring *ring, struct cdnsp_segment *first, struct cdnsp_segment *last, unsigned int num_segs) argument
193 cdnsp_insert_segment_mapping(struct radix_tree_root *trb_address_map, struct cdnsp_ring *ring, struct cdnsp_segment *seg, gfp_t mem_flags) argument
227 cdnsp_update_stream_segment_mapping(struct radix_tree_root *trb_address_map, struct cdnsp_ring *ring, struct cdnsp_segment *first_seg, struct cdnsp_segment *last_seg, gfp_t mem_flags) argument
263 cdnsp_remove_stream_mapping(struct cdnsp_ring *ring) argument
274 cdnsp_update_stream_mapping(struct cdnsp_ring *ring) argument
280 cdnsp_ring_free(struct cdnsp_device *pdev, struct cdnsp_ring *ring) argument
297 cdnsp_initialize_ring_info(struct cdnsp_ring *ring) argument
376 struct cdnsp_ring *ring; local
421 cdnsp_ring_expansion(struct cdnsp_device *pdev, struct cdnsp_ring *ring, unsigned int num_trbs, gfp_t flags) argument
[all...]
/linux-master/tools/include/io_uring/
H A Dmini_liburing.h130 struct io_uring *ring,
136 memset(ring, 0, sizeof(*ring));
143 ret = io_uring_mmap(fd, &p, &ring->sq, &ring->cq);
145 ring->ring_fd = fd;
152 static inline struct io_uring_sqe *io_uring_get_sqe(struct io_uring *ring) argument
154 struct io_uring_sq *sq = &ring->sq;
161 static inline int io_uring_wait_cqe(struct io_uring *ring, argument
164 struct io_uring_cq *cq = &ring
129 io_uring_queue_init(unsigned int entries, struct io_uring *ring, unsigned int flags) argument
185 io_uring_submit(struct io_uring *ring) argument
220 io_uring_queue_exit(struct io_uring *ring) argument
247 io_uring_register_buffers(struct io_uring *ring, const struct iovec *iovecs, unsigned int nr_iovecs) argument
278 io_uring_cqe_seen(struct io_uring *ring) argument
[all...]
/linux-master/drivers/gpu/drm/msm/adreno/
H A Da5xx_gpu.c21 static void update_shadow_rptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument
27 OUT_PKT7(ring, CP_WHERE_AM_I, 2);
28 OUT_RING(ring, lower_32_bits(shadowptr(a5xx_gpu, ring)));
29 OUT_RING(ring, upper_32_bits(shadowptr(a5xx_gpu, ring)));
33 void a5xx_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring, argument
46 update_shadow_rptr(gpu, ring);
48 spin_lock_irqsave(&ring->preempt_lock, flags);
51 ring
68 struct msm_ringbuffer *ring = submit->ring; local
128 struct msm_ringbuffer *ring = submit->ring; local
469 struct msm_ringbuffer *ring = gpu->rb[0]; local
511 struct msm_ringbuffer *ring = gpu->rb[0]; local
1066 a5xx_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument
1221 struct msm_ringbuffer *ring = gpu->funcs->active_ring(gpu); local
1668 a5xx_get_rptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument
[all...]
H A Dadreno_gpu.c569 struct msm_ringbuffer *ring = gpu->rb[i]; local
571 if (!ring)
574 ring->cur = ring->start;
575 ring->next = ring->start;
576 ring->memptrs->rptr = 0;
577 ring->memptrs->bv_fence = ring->fctx->completed_fence;
583 if (fence_before(ring
592 get_rptr(struct adreno_gpu *adreno_gpu, struct msm_ringbuffer *ring) argument
623 adreno_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring, u32 reg) argument
643 adreno_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument
925 struct msm_ringbuffer *ring = gpu->rb[i]; local
959 ring_freewords(struct msm_ringbuffer *ring) argument
969 adreno_wait_ring(struct msm_ringbuffer *ring, uint32_t ndwords) argument
[all...]
/linux-master/drivers/net/ethernet/apm/xgene-v2/
H A DMakefile6 xgene-enet-v2-objs := main.o mac.o enet.o ring.o mdio.o ethtool.o
/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dadf_transport_internal.h22 spinlock_t lock; /* protects ring data struct */
55 int adf_ring_debugfs_add(struct adf_etr_ring_data *ring, const char *name);
56 void adf_ring_debugfs_rm(struct adf_etr_ring_data *ring);
65 static inline int adf_ring_debugfs_add(struct adf_etr_ring_data *ring, argument
71 #define adf_ring_debugfs_rm(ring) do {} while (0)
/linux-master/drivers/gpu/drm/xe/
H A Dxe_lrc_types.h14 * struct xe_lrc - Logical ring context (LRC) and submission ring object
18 * @bo: buffer object (memory) for logical ring context, per process HW
19 * status page, and submission ring.
29 /** @ring: submission ring state */
31 /** @ring.size: size of submission ring */
33 /** @ring.tail: tail of submission ring */
37 } ring; member in struct:xe_lrc
[all...]
/linux-master/virt/kvm/
H A Dcoalesced_mmio.c45 struct kvm_coalesced_mmio_ring *ring; local
54 ring = dev->kvm->coalesced_mmio_ring;
55 avail = (ring->first - last - 1) % KVM_COALESCED_MMIO_MAX;
69 struct kvm_coalesced_mmio_ring *ring = dev->kvm->coalesced_mmio_ring; local
77 insert = READ_ONCE(ring->last);
84 /* copy data in first free entry of the ring */
86 ring->coalesced_mmio[insert].phys_addr = addr;
87 ring->coalesced_mmio[insert].len = len;
88 memcpy(ring->coalesced_mmio[insert].data, val, len);
89 ring
[all...]
/linux-master/drivers/gpu/drm/radeon/
H A Dradeon_vce.c339 * @ring: ring we should submit the msg to
345 int radeon_vce_get_create_msg(struct radeon_device *rdev, int ring, argument
353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4);
406 * @ring: ring we should submit the msg to
412 int radeon_vce_get_destroy_msg(struct radeon_device *rdev, int ring, argument
420 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4);
689 * @ring: engine to use
695 struct radeon_ring *ring,
694 radeon_vce_semaphore_emit(struct radeon_device *rdev, struct radeon_ring *ring, struct radeon_semaphore *semaphore, bool emit_wait) argument
720 struct radeon_ring *ring = &rdev->ring[ib->ring]; local
737 struct radeon_ring *ring = &rdev->ring[fence->ring]; local
755 radeon_vce_ring_test(struct radeon_device *rdev, struct radeon_ring *ring) argument
795 radeon_vce_ib_test(struct radeon_device *rdev, struct radeon_ring *ring) argument
[all...]
/linux-master/drivers/net/ethernet/cisco/enic/
H A Dvnic_cq.c18 vnic_dev_free_desc_ring(cq->vdev, &cq->ring);
35 return vnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size);
46 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET;
48 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size);
72 vnic_dev_clear_desc_ring(&cq->ring);
/linux-master/drivers/scsi/fnic/
H A Dvnic_cq.c14 vnic_dev_free_desc_ring(cq->vdev, &cq->ring);
33 err = vnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size);
48 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET;
50 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size);
72 vnic_dev_clear_desc_ring(&cq->ring);
/linux-master/drivers/scsi/snic/
H A Dvnic_cq.c12 svnic_dev_free_desc_ring(cq->vdev, &cq->ring);
30 return svnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size);
41 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET;
43 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size);
65 svnic_dev_clear_desc_ring(&cq->ring);
H A Dvnic_dev.c176 unsigned int svnic_dev_desc_ring_size(struct vnic_dev_ring *ring, argument
189 ring->base_align = 512;
194 ring->desc_count = ALIGN(desc_count, count_align);
196 ring->desc_size = ALIGN(desc_size, desc_align);
198 ring->size = ring->desc_count * ring->desc_size;
199 ring->size_unaligned = ring->size + ring
204 svnic_dev_clear_desc_ring(struct vnic_dev_ring *ring) argument
209 svnic_dev_alloc_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring, unsigned int desc_count, unsigned int desc_size) argument
236 svnic_dev_free_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring) argument
[all...]
/linux-master/drivers/net/wireless/broadcom/b43/
H A Ddma.h199 struct b43_dmadesc_generic *(*idx2desc) (struct b43_dmaring * ring,
203 void (*fill_descriptor) (struct b43_dmaring * ring,
207 void (*poke_tx) (struct b43_dmaring * ring, int slot);
208 void (*tx_suspend) (struct b43_dmaring * ring);
209 void (*tx_resume) (struct b43_dmaring * ring);
210 int (*get_current_rxslot) (struct b43_dmaring * ring);
211 void (*set_current_rxslot) (struct b43_dmaring * ring, int slot);
229 /* Kernel virtual base address of the ring memory. */
235 * This is NULL for an RX ring.
238 /* (Unadjusted) DMA base bus-address of the ring memor
278 b43_dma_read(struct b43_dmaring *ring, u16 offset) argument
283 b43_dma_write(struct b43_dmaring *ring, u16 offset, u32 value) argument
[all...]
/linux-master/tools/testing/vsock/
H A Dvsock_uring_test.c64 struct io_uring ring; local
80 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring, 0))
83 if (io_uring_register_buffers(&ring, iovec, test_data->vecs_cnt))
89 sqe = io_uring_get_sqe(&ring);
96 if (io_uring_submit(&ring) != 1)
99 if (io_uring_wait_cqe(&ring, &cqe))
102 io_uring_cqe_seen(&ring, cqe);
107 io_uring_queue_exit(&ring);
117 struct io_uring ring; local
137 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring,
[all...]
/linux-master/drivers/net/ethernet/apm/xgene/
H A Dxgene_enet_main.c390 static __le64 *xgene_enet_get_exp_bufs(struct xgene_enet_desc_ring *ring) argument
394 exp_bufs = &ring->exp_bufs[ring->exp_buf_tail * MAX_EXP_BUFFS];
396 ring->exp_buf_tail = (ring->exp_buf_tail + 1) & ((ring->slots / 2) - 1);
401 static dma_addr_t *xgene_get_frag_dma_array(struct xgene_enet_desc_ring *ring) argument
403 return &ring->cp_ring->frag_dma_addr[ring->tail * MAX_SKB_FRAGS];
783 static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring, argument
850 struct xgene_enet_desc_ring *ring; local
882 struct xgene_enet_desc_ring *ring; local
907 struct xgene_enet_desc_ring *ring; local
939 struct xgene_enet_desc_ring *ring; local
1048 xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring) argument
1063 struct xgene_enet_desc_ring *ring; local
1129 xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring) argument
1151 struct xgene_enet_desc_ring *ring; local
1195 is_irq_mbox_required(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument
1206 xgene_enet_ring_cmd_base(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument
1220 struct xgene_enet_desc_ring *ring; local
1477 struct xgene_enet_desc_ring *ring; local
[all...]

Completed in 225 milliseconds

1234567891011>>