/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | jpeg_v3_0.c | 47 * Set ring and irq function pointers 85 struct amdgpu_ring *ring; local 102 ring = adev->jpeg.inst->ring_dec; 103 ring->use_doorbell = true; 104 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1; 105 ring->vm_hub = AMDGPU_MMHUB0(0); 106 sprintf(ring->name, "jpeg_dec"); 107 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst->irq, 0, 148 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local 151 adev->nbio.funcs->vcn_doorbell_range(adev, ring 334 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local 414 jpeg_v3_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 428 jpeg_v3_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 445 jpeg_v3_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument [all...] |
H A D | sdma_v4_0.c | 602 * @ring: amdgpu ring pointer 606 static uint64_t sdma_v4_0_ring_get_rptr(struct amdgpu_ring *ring) argument 611 rptr = ((u64 *)ring->rptr_cpu_addr); 620 * @ring: amdgpu ring pointer 624 static uint64_t sdma_v4_0_ring_get_wptr(struct amdgpu_ring *ring) argument 626 struct amdgpu_device *adev = ring->adev; 629 if (ring->use_doorbell) { 631 wptr = READ_ONCE(*((u64 *)ring 651 sdma_v4_0_ring_set_wptr(struct amdgpu_ring *ring) argument 693 sdma_v4_0_page_ring_get_wptr(struct amdgpu_ring *ring) argument 717 sdma_v4_0_page_ring_set_wptr(struct amdgpu_ring *ring) argument 737 sdma_v4_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 760 sdma_v4_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 781 sdma_v4_0_wait_reg_mem(struct amdgpu_ring *ring, int mem_space, int hdp, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 813 sdma_v4_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 839 sdma_v4_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 1019 sdma_v4_0_rb_cntl(struct amdgpu_ring *ring, uint32_t rb_cntl) argument 1044 struct amdgpu_ring *ring = &adev->sdma.instance[i].ring; local 1129 struct amdgpu_ring *ring = &adev->sdma.instance[i].page; local 1337 struct amdgpu_ring *ring; local 1416 sdma_v4_0_ring_test_ring(struct amdgpu_ring *ring) argument 1469 sdma_v4_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1621 sdma_v4_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1646 sdma_v4_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1669 sdma_v4_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1675 sdma_v4_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1684 sdma_v4_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1750 struct amdgpu_ring *ring; local [all...] |
H A D | gfx_v11_0.c | 133 static void gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume); 134 static void gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure); 135 static void gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, 138 static void gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, 161 struct amdgpu_ring *ring) 163 uint64_t mqd_addr = amdgpu_bo_gpu_offset(ring->mqd_obj); 164 uint64_t wptr_addr = ring->wptr_gpu_addr; 167 switch (ring->funcs->type) { 189 PACKET3_MAP_QUEUES_QUEUE(ring->queue) | 190 PACKET3_MAP_QUEUES_PIPE(ring 160 gfx11_kiq_map_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring) argument 203 gfx11_kiq_unmap_queues(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, enum amdgpu_unmap_queues_action action, u64 gpu_addr, u64 seq) argument 236 gfx11_kiq_query_status(struct amdgpu_ring *kiq_ring, struct amdgpu_ring *ring, u64 addr, u64 seq) argument 303 gfx_v11_0_write_data_to_reg(struct amdgpu_ring *ring, int eng_sel, bool wc, uint32_t reg, uint32_t val) argument 314 gfx_v11_0_wait_reg_mem(struct amdgpu_ring *ring, int eng_sel, int mem_space, int opt, uint32_t addr0, uint32_t addr1, uint32_t ref, uint32_t mask, uint32_t inv) argument 336 gfx_v11_0_ring_test_ring(struct amdgpu_ring *ring) argument 377 gfx_v11_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 933 struct amdgpu_ring *ring; local 965 struct amdgpu_ring *ring; local 3151 struct amdgpu_ring *ring; local 3236 gfx_v11_0_cp_gfx_set_doorbell(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 3263 struct amdgpu_ring *ring; local 3589 gfx_v11_0_kiq_setting(struct amdgpu_ring *ring) argument 3704 gfx_v11_0_gfx_init_queue(struct amdgpu_ring *ring) argument 3735 struct amdgpu_ring *ring; local 3892 gfx_v11_0_kiq_init_register(struct amdgpu_ring *ring) argument 4002 gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring) argument 4041 gfx_v11_0_kcq_init_queue(struct amdgpu_ring *ring) argument 4072 struct amdgpu_ring *ring; local 4097 struct amdgpu_ring *ring = NULL; local 4128 struct amdgpu_ring *ring; local 4630 struct amdgpu_ring *ring; local 4687 gfx_v11_0_ring_emit_gds_switch(struct amdgpu_ring *ring, uint32_t vmid, uint32_t gds_base, uint32_t gds_size, uint32_t gws_base, uint32_t gws_size, uint32_t oa_base, uint32_t oa_size) argument 5036 gfx_v11_0_update_spm_vmid(struct amdgpu_device *adev, struct amdgpu_ring *ring, unsigned vmid) argument 5211 gfx_v11_0_ring_get_rptr_gfx(struct amdgpu_ring *ring) argument 5217 gfx_v11_0_ring_get_wptr_gfx(struct amdgpu_ring *ring) argument 5233 gfx_v11_0_ring_set_wptr_gfx(struct amdgpu_ring *ring) argument 5250 gfx_v11_0_ring_get_rptr_compute(struct amdgpu_ring *ring) argument 5256 gfx_v11_0_ring_get_wptr_compute(struct amdgpu_ring *ring) argument 5268 gfx_v11_0_ring_set_wptr_compute(struct amdgpu_ring *ring) argument 5282 gfx_v11_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 5311 gfx_v11_0_ring_emit_ib_gfx(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5351 gfx_v11_0_ring_emit_ib_compute(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 5390 gfx_v11_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 5428 gfx_v11_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 5438 gfx_v11_0_ring_invalidate_tlbs(struct amdgpu_ring *ring, uint16_t pasid, uint32_t flush_type, bool all_hub, uint8_t dst_sel) argument 5450 gfx_v11_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 5472 gfx_v11_0_ring_emit_fence_kiq(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 5499 gfx_v11_0_ring_emit_cntxcntl(struct amdgpu_ring *ring, uint32_t flags) argument 5519 gfx_v11_0_ring_emit_init_cond_exec(struct amdgpu_ring *ring, uint64_t addr) argument 5536 gfx_v11_0_ring_emit_gfx_shadow(struct amdgpu_ring *ring, u64 shadow_va, u64 csa_va, u64 gds_va, bool init_shadow, int vmid) argument 5627 gfx_v11_0_ring_preempt_ib(struct amdgpu_ring *ring) argument 5674 gfx_v11_0_ring_emit_de_meta(struct amdgpu_ring *ring, bool resume) argument 5725 gfx_v11_0_ring_emit_frame_cntl(struct amdgpu_ring *ring, bool start, bool secure) argument 5734 gfx_v11_0_ring_emit_rreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t reg_val_offs) argument 5751 gfx_v11_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 5774 gfx_v11_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 5780 gfx_v11_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 5790 gfx_v11_0_ring_soft_recovery(struct amdgpu_ring *ring, unsigned vmid) argument 5942 struct amdgpu_ring *ring; local 6033 struct amdgpu_ring *ring; local 6136 gfx_v11_0_emit_mem_sync(struct amdgpu_ring *ring) argument [all...] |
H A D | jpeg_v4_0_5.c | 54 static void jpeg_v4_0_5_dec_ring_set_wptr(struct amdgpu_ring *ring); 66 * Set ring and irq function pointers 104 struct amdgpu_ring *ring; local 142 ring = adev->jpeg.inst[i].ring_dec; 143 ring->use_doorbell = true; 144 ring->vm_hub = AMDGPU_MMHUB0(0); 145 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1 + 8 * i; 146 sprintf(ring->name, "jpeg_dec_%d", i); 147 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst[i].irq, 189 struct amdgpu_ring *ring; local 399 struct amdgpu_ring *ring = adev->jpeg.inst[inst_idx].ring_dec; local 479 struct amdgpu_ring *ring; local 585 jpeg_v4_0_5_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 599 jpeg_v4_0_5_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 616 jpeg_v4_0_5_dec_ring_set_wptr(struct amdgpu_ring *ring) argument [all...] |
H A D | psp_v10_0.c | 75 struct psp_ring *ring = &psp->km_ring; local 78 /* Write low address of the ring to C2PMSG_69 */ 79 psp_ring_reg = lower_32_bits(ring->ring_mem_mc_addr); 81 /* Write high address of the ring to C2PMSG_70 */ 82 psp_ring_reg = upper_32_bits(ring->ring_mem_mc_addr); 84 /* Write size of ring to C2PMSG_71 */ 85 psp_ring_reg = ring->ring_size; 87 /* Write the ring initialization command to C2PMSG_64 */ 109 /* Write the ring destroy command to C2PMSG_64 */ 127 struct psp_ring *ring local [all...] |
H A D | jpeg_v4_0.c | 46 static void jpeg_v4_0_dec_ring_set_wptr(struct amdgpu_ring *ring); 53 * Set ring and irq function pointers 80 struct amdgpu_ring *ring; local 109 ring = adev->jpeg.inst->ring_dec; 110 ring->use_doorbell = true; 111 ring->doorbell_index = amdgpu_sriov_vf(adev) ? (((adev->doorbell_index.vcn.vcn_ring0_1) << 1) + 4) : ((adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 1); 112 ring->vm_hub = AMDGPU_MMHUB0(0); 114 sprintf(ring->name, "jpeg_dec"); 115 r = amdgpu_ring_init(adev, ring, 512, &adev->jpeg.inst->irq, 0, 160 struct amdgpu_ring *ring local 368 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; local 413 struct amdgpu_ring *ring; local 571 jpeg_v4_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 585 jpeg_v4_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 602 jpeg_v4_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_ids.c | 190 * @ring: ring we want to submit job to 197 static int amdgpu_vmid_grab_idle(struct amdgpu_ring *ring, argument 201 struct amdgpu_device *adev = ring->adev; 202 unsigned vmhub = ring->vm_hub; 207 if (!dma_fence_is_signaled(ring->vmid_wait)) { 208 *fence = dma_fence_get(ring->vmid_wait); 221 NULL : ring; 231 u64 fence_context = adev->vm_manager.fence_context + ring->idx; 232 unsigned seqno = ++adev->vm_manager.seqno[ring 270 amdgpu_vmid_grab_reserved(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_vmid **id, struct dma_fence **fence) argument 331 amdgpu_vmid_grab_used(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_vmid **id, struct dma_fence **fence) argument 394 amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct dma_fence **fence) argument [all...] |
H A D | amdgpu_vcn.c | 280 static bool amdgpu_vcn_using_unified_queue(struct amdgpu_ring *ring) argument 282 struct amdgpu_device *adev = ring->adev; 429 void amdgpu_vcn_ring_begin_use(struct amdgpu_ring *ring) argument 431 struct amdgpu_device *adev = ring->adev; 450 if (ring->funcs->type == AMDGPU_RING_TYPE_VCN_ENC) { 451 atomic_inc(&adev->vcn.inst[ring->me].dpg_enc_submission_cnt); 458 fences += amdgpu_fence_count_emitted(&adev->vcn.inst[ring->me].ring_enc[i]); 460 if (fences || atomic_read(&adev->vcn.inst[ring->me].dpg_enc_submission_cnt)) 466 adev->vcn.pause_dpg_mode(adev, ring->me, &new_state); 471 void amdgpu_vcn_ring_end_use(struct amdgpu_ring *ring) argument 482 amdgpu_vcn_dec_ring_test_ring(struct amdgpu_ring *ring) argument 513 amdgpu_vcn_dec_sw_ring_test_ring(struct amdgpu_ring *ring) argument 544 amdgpu_vcn_dec_send_msg(struct amdgpu_ring *ring, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument 593 amdgpu_vcn_dec_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) argument 628 amdgpu_vcn_dec_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) argument 655 amdgpu_vcn_dec_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 717 amdgpu_vcn_dec_sw_send_msg(struct amdgpu_ring *ring, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument 787 amdgpu_vcn_dec_sw_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 819 amdgpu_vcn_enc_ring_test_ring(struct amdgpu_ring *ring) argument 850 amdgpu_vcn_enc_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument 917 amdgpu_vcn_enc_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib_msg, struct dma_fence **fence) argument 984 amdgpu_vcn_enc_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1019 amdgpu_vcn_unified_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1036 amdgpu_vcn_get_enc_ring_prio(int ring) argument [all...] |
/linux-master/drivers/net/ethernet/broadcom/genet/ |
H A D | bcmgenet.c | 345 /* RDMA/TDMA ring registers and accessors 409 unsigned int ring, 413 (DMA_RING_SIZE * ring) + 418 unsigned int ring, u32 val, 422 (DMA_RING_SIZE * ring) + 427 unsigned int ring, 431 (DMA_RING_SIZE * ring) + 436 unsigned int ring, u32 val, 440 (DMA_RING_SIZE * ring) + 672 * We'll map them to ring 408 bcmgenet_tdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument 417 bcmgenet_tdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument 426 bcmgenet_rdma_ring_readl(struct bcmgenet_priv *priv, unsigned int ring, enum dma_ring_reg r) argument 435 bcmgenet_rdma_ring_writel(struct bcmgenet_priv *priv, unsigned int ring, u32 val, enum dma_ring_reg r) argument 818 struct bcmgenet_rx_ring *ring; local 840 bcmgenet_set_rx_coalesce(struct bcmgenet_rx_ring *ring, u32 usecs, u32 pkts) argument 855 bcmgenet_set_ring_rx_coalesce(struct bcmgenet_rx_ring *ring, struct ethtool_coalesce *ec) argument 1728 bcmgenet_get_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument 1745 bcmgenet_put_txcb(struct bcmgenet_priv *priv, struct bcmgenet_tx_ring *ring) argument 1762 bcmgenet_rx_ring16_int_disable(struct bcmgenet_rx_ring *ring) argument 1768 bcmgenet_rx_ring16_int_enable(struct bcmgenet_rx_ring *ring) argument 1774 bcmgenet_rx_ring_int_disable(struct bcmgenet_rx_ring *ring) argument 1781 bcmgenet_rx_ring_int_enable(struct bcmgenet_rx_ring *ring) argument 1788 bcmgenet_tx_ring16_int_disable(struct bcmgenet_tx_ring *ring) argument 1794 bcmgenet_tx_ring16_int_enable(struct bcmgenet_tx_ring *ring) argument 1800 bcmgenet_tx_ring_int_enable(struct bcmgenet_tx_ring *ring) argument 1806 bcmgenet_tx_ring_int_disable(struct bcmgenet_tx_ring *ring) argument 1868 __bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument 1925 bcmgenet_tx_reclaim(struct net_device *dev, struct bcmgenet_tx_ring *ring) argument 1939 struct bcmgenet_tx_ring *ring = local 2047 struct bcmgenet_tx_ring *ring = NULL; local 2229 bcmgenet_desc_rx(struct bcmgenet_rx_ring *ring, unsigned int budget) argument 2393 struct bcmgenet_rx_ring *ring = container_of(napi, local 2419 struct bcmgenet_rx_ring *ring = local 2429 bcmgenet_alloc_rx_buffers(struct bcmgenet_priv *priv, struct bcmgenet_rx_ring *ring) argument 2599 bcmgenet_init_dim(struct bcmgenet_rx_ring *ring, void (*cb)(struct work_struct *work)) argument 2611 bcmgenet_init_rx_coalesce(struct bcmgenet_rx_ring *ring) argument 2635 struct bcmgenet_tx_ring *ring = &priv->tx_rings[index]; local 2694 struct bcmgenet_rx_ring *ring = &priv->rx_rings[index]; local 2750 struct bcmgenet_tx_ring *ring; local 2766 struct bcmgenet_tx_ring *ring; local 2780 struct bcmgenet_tx_ring *ring; local 2863 struct bcmgenet_rx_ring *ring; local 2879 struct bcmgenet_rx_ring *ring; local 2895 struct bcmgenet_rx_ring *ring; local 3498 bcmgenet_dump_tx_queue(struct bcmgenet_tx_ring *ring) argument [all...] |
/linux-master/drivers/usb/cdns3/ |
H A D | cdnsp-mem.c | 24 * Allocates a generic ring segment from the ring pool, sets the dma address, 133 * Link the ring to the new segments. 134 * Set Toggle Cycle for the new ring if needed. 137 struct cdnsp_ring *ring, 144 if (!ring || !first || !last) 147 next = ring->enq_seg->next; 148 cdnsp_link_segments(pdev, ring->enq_seg, first, ring->type); 149 cdnsp_link_segments(pdev, last, next, ring 136 cdnsp_link_rings(struct cdnsp_device *pdev, struct cdnsp_ring *ring, struct cdnsp_segment *first, struct cdnsp_segment *last, unsigned int num_segs) argument 193 cdnsp_insert_segment_mapping(struct radix_tree_root *trb_address_map, struct cdnsp_ring *ring, struct cdnsp_segment *seg, gfp_t mem_flags) argument 227 cdnsp_update_stream_segment_mapping(struct radix_tree_root *trb_address_map, struct cdnsp_ring *ring, struct cdnsp_segment *first_seg, struct cdnsp_segment *last_seg, gfp_t mem_flags) argument 263 cdnsp_remove_stream_mapping(struct cdnsp_ring *ring) argument 274 cdnsp_update_stream_mapping(struct cdnsp_ring *ring) argument 280 cdnsp_ring_free(struct cdnsp_device *pdev, struct cdnsp_ring *ring) argument 297 cdnsp_initialize_ring_info(struct cdnsp_ring *ring) argument 376 struct cdnsp_ring *ring; local 421 cdnsp_ring_expansion(struct cdnsp_device *pdev, struct cdnsp_ring *ring, unsigned int num_trbs, gfp_t flags) argument [all...] |
/linux-master/tools/include/io_uring/ |
H A D | mini_liburing.h | 130 struct io_uring *ring, 136 memset(ring, 0, sizeof(*ring)); 143 ret = io_uring_mmap(fd, &p, &ring->sq, &ring->cq); 145 ring->ring_fd = fd; 152 static inline struct io_uring_sqe *io_uring_get_sqe(struct io_uring *ring) argument 154 struct io_uring_sq *sq = &ring->sq; 161 static inline int io_uring_wait_cqe(struct io_uring *ring, argument 164 struct io_uring_cq *cq = &ring 129 io_uring_queue_init(unsigned int entries, struct io_uring *ring, unsigned int flags) argument 185 io_uring_submit(struct io_uring *ring) argument 220 io_uring_queue_exit(struct io_uring *ring) argument 247 io_uring_register_buffers(struct io_uring *ring, const struct iovec *iovecs, unsigned int nr_iovecs) argument 278 io_uring_cqe_seen(struct io_uring *ring) argument [all...] |
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a5xx_gpu.c | 21 static void update_shadow_rptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument 27 OUT_PKT7(ring, CP_WHERE_AM_I, 2); 28 OUT_RING(ring, lower_32_bits(shadowptr(a5xx_gpu, ring))); 29 OUT_RING(ring, upper_32_bits(shadowptr(a5xx_gpu, ring))); 33 void a5xx_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring, argument 46 update_shadow_rptr(gpu, ring); 48 spin_lock_irqsave(&ring->preempt_lock, flags); 51 ring 68 struct msm_ringbuffer *ring = submit->ring; local 128 struct msm_ringbuffer *ring = submit->ring; local 469 struct msm_ringbuffer *ring = gpu->rb[0]; local 511 struct msm_ringbuffer *ring = gpu->rb[0]; local 1066 a5xx_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument 1221 struct msm_ringbuffer *ring = gpu->funcs->active_ring(gpu); local 1668 a5xx_get_rptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument [all...] |
H A D | adreno_gpu.c | 569 struct msm_ringbuffer *ring = gpu->rb[i]; local 571 if (!ring) 574 ring->cur = ring->start; 575 ring->next = ring->start; 576 ring->memptrs->rptr = 0; 577 ring->memptrs->bv_fence = ring->fctx->completed_fence; 583 if (fence_before(ring 592 get_rptr(struct adreno_gpu *adreno_gpu, struct msm_ringbuffer *ring) argument 623 adreno_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring, u32 reg) argument 643 adreno_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument 925 struct msm_ringbuffer *ring = gpu->rb[i]; local 959 ring_freewords(struct msm_ringbuffer *ring) argument 969 adreno_wait_ring(struct msm_ringbuffer *ring, uint32_t ndwords) argument [all...] |
/linux-master/drivers/net/ethernet/apm/xgene-v2/ |
H A D | Makefile | 6 xgene-enet-v2-objs := main.o mac.o enet.o ring.o mdio.o ethtool.o
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_transport_internal.h | 22 spinlock_t lock; /* protects ring data struct */ 55 int adf_ring_debugfs_add(struct adf_etr_ring_data *ring, const char *name); 56 void adf_ring_debugfs_rm(struct adf_etr_ring_data *ring); 65 static inline int adf_ring_debugfs_add(struct adf_etr_ring_data *ring, argument 71 #define adf_ring_debugfs_rm(ring) do {} while (0)
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_lrc_types.h | 14 * struct xe_lrc - Logical ring context (LRC) and submission ring object 18 * @bo: buffer object (memory) for logical ring context, per process HW 19 * status page, and submission ring. 29 /** @ring: submission ring state */ 31 /** @ring.size: size of submission ring */ 33 /** @ring.tail: tail of submission ring */ 37 } ring; member in struct:xe_lrc [all...] |
/linux-master/virt/kvm/ |
H A D | coalesced_mmio.c | 45 struct kvm_coalesced_mmio_ring *ring; local 54 ring = dev->kvm->coalesced_mmio_ring; 55 avail = (ring->first - last - 1) % KVM_COALESCED_MMIO_MAX; 69 struct kvm_coalesced_mmio_ring *ring = dev->kvm->coalesced_mmio_ring; local 77 insert = READ_ONCE(ring->last); 84 /* copy data in first free entry of the ring */ 86 ring->coalesced_mmio[insert].phys_addr = addr; 87 ring->coalesced_mmio[insert].len = len; 88 memcpy(ring->coalesced_mmio[insert].data, val, len); 89 ring [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_vce.c | 339 * @ring: ring we should submit the msg to 345 int radeon_vce_get_create_msg(struct radeon_device *rdev, int ring, argument 353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); 406 * @ring: ring we should submit the msg to 412 int radeon_vce_get_destroy_msg(struct radeon_device *rdev, int ring, argument 420 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); 689 * @ring: engine to use 695 struct radeon_ring *ring, 694 radeon_vce_semaphore_emit(struct radeon_device *rdev, struct radeon_ring *ring, struct radeon_semaphore *semaphore, bool emit_wait) argument 720 struct radeon_ring *ring = &rdev->ring[ib->ring]; local 737 struct radeon_ring *ring = &rdev->ring[fence->ring]; local 755 radeon_vce_ring_test(struct radeon_device *rdev, struct radeon_ring *ring) argument 795 radeon_vce_ib_test(struct radeon_device *rdev, struct radeon_ring *ring) argument [all...] |
/linux-master/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_cq.c | 18 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); 35 return vnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size); 46 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET; 48 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size); 72 vnic_dev_clear_desc_ring(&cq->ring);
|
/linux-master/drivers/scsi/fnic/ |
H A D | vnic_cq.c | 14 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); 33 err = vnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size); 48 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET; 50 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size); 72 vnic_dev_clear_desc_ring(&cq->ring);
|
/linux-master/drivers/scsi/snic/ |
H A D | vnic_cq.c | 12 svnic_dev_free_desc_ring(cq->vdev, &cq->ring); 30 return svnic_dev_alloc_desc_ring(vdev, &cq->ring, desc_count, desc_size); 41 paddr = (u64)cq->ring.base_addr | VNIC_PADDR_TARGET; 43 iowrite32(cq->ring.desc_count, &cq->ctrl->ring_size); 65 svnic_dev_clear_desc_ring(&cq->ring);
|
H A D | vnic_dev.c | 176 unsigned int svnic_dev_desc_ring_size(struct vnic_dev_ring *ring, argument 189 ring->base_align = 512; 194 ring->desc_count = ALIGN(desc_count, count_align); 196 ring->desc_size = ALIGN(desc_size, desc_align); 198 ring->size = ring->desc_count * ring->desc_size; 199 ring->size_unaligned = ring->size + ring 204 svnic_dev_clear_desc_ring(struct vnic_dev_ring *ring) argument 209 svnic_dev_alloc_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring, unsigned int desc_count, unsigned int desc_size) argument 236 svnic_dev_free_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring) argument [all...] |
/linux-master/drivers/net/wireless/broadcom/b43/ |
H A D | dma.h | 199 struct b43_dmadesc_generic *(*idx2desc) (struct b43_dmaring * ring, 203 void (*fill_descriptor) (struct b43_dmaring * ring, 207 void (*poke_tx) (struct b43_dmaring * ring, int slot); 208 void (*tx_suspend) (struct b43_dmaring * ring); 209 void (*tx_resume) (struct b43_dmaring * ring); 210 int (*get_current_rxslot) (struct b43_dmaring * ring); 211 void (*set_current_rxslot) (struct b43_dmaring * ring, int slot); 229 /* Kernel virtual base address of the ring memory. */ 235 * This is NULL for an RX ring. 238 /* (Unadjusted) DMA base bus-address of the ring memor 278 b43_dma_read(struct b43_dmaring *ring, u16 offset) argument 283 b43_dma_write(struct b43_dmaring *ring, u16 offset, u32 value) argument [all...] |
/linux-master/tools/testing/vsock/ |
H A D | vsock_uring_test.c | 64 struct io_uring ring; local 80 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring, 0)) 83 if (io_uring_register_buffers(&ring, iovec, test_data->vecs_cnt)) 89 sqe = io_uring_get_sqe(&ring); 96 if (io_uring_submit(&ring) != 1) 99 if (io_uring_wait_cqe(&ring, &cqe)) 102 io_uring_cqe_seen(&ring, cqe); 107 io_uring_queue_exit(&ring); 117 struct io_uring ring; local 137 if (io_uring_queue_init(RING_ENTRIES_NUM, &ring, [all...] |
/linux-master/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_main.c | 390 static __le64 *xgene_enet_get_exp_bufs(struct xgene_enet_desc_ring *ring) argument 394 exp_bufs = &ring->exp_bufs[ring->exp_buf_tail * MAX_EXP_BUFFS]; 396 ring->exp_buf_tail = (ring->exp_buf_tail + 1) & ((ring->slots / 2) - 1); 401 static dma_addr_t *xgene_get_frag_dma_array(struct xgene_enet_desc_ring *ring) argument 403 return &ring->cp_ring->frag_dma_addr[ring->tail * MAX_SKB_FRAGS]; 783 static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring, argument 850 struct xgene_enet_desc_ring *ring; local 882 struct xgene_enet_desc_ring *ring; local 907 struct xgene_enet_desc_ring *ring; local 939 struct xgene_enet_desc_ring *ring; local 1048 xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring) argument 1063 struct xgene_enet_desc_ring *ring; local 1129 xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring) argument 1151 struct xgene_enet_desc_ring *ring; local 1195 is_irq_mbox_required(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument 1206 xgene_enet_ring_cmd_base(struct xgene_enet_pdata *pdata, struct xgene_enet_desc_ring *ring) argument 1220 struct xgene_enet_desc_ring *ring; local 1477 struct xgene_enet_desc_ring *ring; local [all...] |