/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_rx.c | 75 struct mlx4_en_rx_ring *ring, 86 ring->rx_alloc_pages++; 109 struct mlx4_en_rx_ring *ring, int index) 111 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index; 122 /* If the number of used fragments does not fill up the ring stride, 125 possible_frags = (ring->stride - sizeof(struct mlx4_en_rx_desc)) / DS_SIZE; 134 struct mlx4_en_rx_ring *ring, int index, 137 struct mlx4_en_rx_desc *rx_desc = ring->buf + 138 (index << ring 74 mlx4_en_alloc_frags(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, gfp_t gfp) argument 108 mlx4_en_init_rx_desc(const struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument 133 mlx4_en_prepare_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index, gfp_t gfp) argument 157 mlx4_en_is_ring_empty(const struct mlx4_en_rx_ring *ring) argument 162 mlx4_en_update_rx_prod_db(struct mlx4_en_rx_ring *ring) argument 168 mlx4_en_free_rx_desc(const struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument 185 struct mlx4_en_rx_ring *ring; local 226 mlx4_en_free_rx_buf(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 268 struct mlx4_en_rx_ring *ring; local 328 struct mlx4_en_rx_ring *ring; local 395 int ring; local 415 mlx4_en_rx_recycle(struct mlx4_en_rx_ring *ring, struct mlx4_en_rx_alloc *frame) argument 434 struct mlx4_en_rx_ring *ring = *pring; local 450 mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 547 mlx4_en_refill_rx_buffers(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 668 struct mlx4_en_rx_ring *ring; member in struct:mlx4_en_xdp_buff 718 struct mlx4_en_rx_ring *ring; local 1114 mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, int qpn, struct mlx4_en_rx_ring *ring, enum mlx4_qp_state *state, struct mlx4_qp *qp) argument [all...] |
/linux-master/drivers/net/ethernet/hisilicon/hns3/ |
H A D | hns3_enet.c | 428 if (tqp_vectors->tx_group.ring && tqp_vectors->rx_group.ring) { 434 } else if (tqp_vectors->rx_group.ring) { 439 } else if (tqp_vectors->tx_group.ring) { 751 struct hns3_enet_ring *ring = tqp_vector->tx_group.ring; local 753 while (ring) { 758 ring->tqp->tqp_index); 763 ring = ring 962 hns3_tx_spare_space(struct hns3_enet_ring *ring) argument 982 hns3_tx_spare_update(struct hns3_enet_ring *ring) argument 997 hns3_can_use_tx_bounce(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 space) argument 1015 hns3_can_use_tx_sgl(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 space) argument 1032 hns3_init_tx_spare_buffer(struct hns3_enet_ring *ring) argument 1089 hns3_tx_spare_alloc(struct hns3_enet_ring *ring, unsigned int size, dma_addr_t *dma, u32 *cb_len) argument 1116 hns3_tx_spare_rollback(struct hns3_enet_ring *ring, u32 len) argument 1128 hns3_tx_spare_reclaim_cb(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument 1579 hns3_handle_vlan_info(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_param *param) argument 1604 hns3_handle_csum_partial(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_cb *desc_cb, struct hns3_desc_param *param) argument 1648 hns3_fill_skb_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc *desc, struct hns3_desc_cb *desc_cb) argument 1680 hns3_fill_desc(struct hns3_enet_ring *ring, dma_addr_t dma, unsigned int size) argument 1723 hns3_map_and_fill_desc(struct hns3_enet_ring *ring, void *priv, unsigned int type) argument 1893 hns3_skb_linearize(struct hns3_enet_ring *ring, struct sk_buff *skb, unsigned int bd_num) argument 1922 hns3_nic_maybe_stop_tx(struct hns3_enet_ring *ring, struct net_device *netdev, struct sk_buff *skb) argument 1970 hns3_clear_desc(struct hns3_enet_ring *ring, int next_to_use_orig) argument 2010 hns3_fill_skb_to_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, unsigned int type) argument 2044 hns3_tx_push_bd(struct hns3_enet_ring *ring, int num) argument 2075 hns3_tx_mem_doorbell(struct hns3_enet_ring *ring) argument 2095 hns3_tx_doorbell(struct hns3_enet_ring *ring, int num, bool doorbell) argument 2148 hns3_handle_tx_bounce(struct hns3_enet_ring *ring, struct sk_buff *skb) argument 2201 hns3_handle_tx_sgl(struct hns3_enet_ring *ring, struct sk_buff *skb) argument 2253 hns3_handle_desc_filling(struct hns3_enet_ring *ring, struct sk_buff *skb) argument 2273 hns3_handle_skb_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_cb *desc_cb, int next_to_use_head) argument 2301 struct hns3_enet_ring *ring = &priv->ring[skb->queue_mapping]; local 2494 hns3_fetch_stats(struct rtnl_link_stats64 *stats, struct hns3_enet_ring *ring, bool is_tx) argument 2544 struct hns3_enet_ring *ring; local 3367 hns3_alloc_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument 3404 hns3_free_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb, int budget) argument 3420 hns3_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument 3431 hns3_unmap_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument 3445 hns3_buffer_detach(struct hns3_enet_ring *ring, int i) argument 3452 hns3_free_buffer_detach(struct hns3_enet_ring *ring, int i, int budget) argument 3464 hns3_free_buffers(struct hns3_enet_ring *ring) argument 3473 hns3_free_desc(struct hns3_enet_ring *ring) argument 3486 hns3_alloc_desc(struct hns3_enet_ring *ring) argument 3498 hns3_alloc_and_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument 3519 hns3_alloc_and_attach_buffer(struct hns3_enet_ring *ring, int i) argument 3534 hns3_alloc_ring_buffers(struct hns3_enet_ring *ring) argument 3553 hns3_replace_buffer(struct hns3_enet_ring *ring, int i, struct hns3_desc_cb *res_cb) argument 3564 hns3_reuse_buffer(struct hns3_enet_ring *ring, int i) argument 3578 hns3_nic_reclaim_desc(struct hns3_enet_ring *ring, int *bytes, int *pkts, int budget) argument 3630 hns3_clean_tx_ring(struct hns3_enet_ring *ring, int budget) argument 3668 hns3_desc_unused(struct hns3_enet_ring *ring) argument 3680 hns3_nic_alloc_rx_buffers(struct hns3_enet_ring *ring, int cleand_count) argument 3723 hns3_handle_rx_copybreak(struct sk_buff *skb, int i, struct hns3_enet_ring *ring, int pull_len, struct hns3_desc_cb *desc_cb) argument 3751 hns3_nic_reuse_page(struct sk_buff *skb, int i, struct hns3_enet_ring *ring, int pull_len, struct hns3_desc_cb *desc_cb) argument 3873 hns3_checksum_complete(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 ptype, u16 csum) argument 3923 hns3_rx_checksum(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 l234info, u32 bd_base_info, u32 ol_info, u16 csum) argument 3960 hns3_rx_skb(struct hns3_enet_ring *ring, struct sk_buff *skb) argument 3968 hns3_parse_vlan_tag(struct hns3_enet_ring *ring, struct hns3_desc *desc, u32 l234info, u16 *vlan_tag) argument 4021 hns3_rx_ring_move_fw(struct hns3_enet_ring *ring) argument 4032 hns3_alloc_skb(struct hns3_enet_ring *ring, unsigned int length, unsigned char *va) argument 4085 hns3_add_frag(struct hns3_enet_ring *ring) argument 4146 hns3_set_gro_and_checksum(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 l234info, u32 bd_base_info, u32 ol_info, u16 csum) argument 4188 hns3_set_rx_skb_rss_type(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 rss_hash, u32 l234info, u32 ol_info) argument 4236 hns3_handle_rx_vlan_tag(struct hns3_enet_ring *ring, struct hns3_desc *desc, struct sk_buff *skb, u32 l234info) argument 4255 hns3_handle_bdinfo(struct hns3_enet_ring *ring, struct sk_buff *skb) argument 4325 hns3_handle_rx_bd(struct hns3_enet_ring *ring) argument 4397 hns3_clean_rx_ring(struct hns3_enet_ring *ring, int budget, void (*rx_fn)(struct hns3_enet_ring *, struct sk_buff *)) argument 4471 struct hns3_enet_ring *ring; local 4528 struct hns3_enet_ring *ring; local 4600 hns3_add_ring_to_group(struct hns3_enet_ring_group *group, struct hns3_enet_ring *ring) argument 4876 struct hns3_enet_ring *ring; local 4940 hns3_alloc_page_pool(struct hns3_enet_ring *ring) argument 4962 hns3_alloc_ring_memory(struct hns3_enet_ring *ring) argument 5002 hns3_fini_ring(struct hns3_enet_ring *ring) argument 5055 hns3_init_ring_hw(struct hns3_enet_ring *ring) argument 5482 hns3_clear_tx_ring(struct hns3_enet_ring *ring) argument 5493 hns3_clear_rx_ring(struct hns3_enet_ring *ring) argument 5530 hns3_force_clear_rx_ring(struct hns3_enet_ring *ring) argument 5554 struct hns3_enet_ring *ring; local [all...] |
/linux-master/drivers/scsi/fnic/ |
H A D | vnic_cq_copy.h | 23 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs + 24 cq->ring.desc_size * cq->to_clean); 33 if (cq->to_clean == cq->ring.desc_count) { 38 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs + 39 cq->ring.desc_size * cq->to_clean);
|
H A D | vnic_wq_copy.h | 19 struct vnic_dev_ring ring; member in struct:vnic_wq_copy 26 return wq->ring.desc_avail; 31 return wq->ring.desc_count - 1 - wq->ring.desc_avail; 36 struct fcpio_host_req *desc = wq->ring.descs; 43 ((wq->to_use_index + 1) == wq->ring.desc_count) ? 45 wq->ring.desc_avail--; 64 cnt = wq->ring.desc_count - wq->to_clean_index + index + 1; 66 wq->to_clean_index = ((index + 1) % wq->ring.desc_count); 67 wq->ring [all...] |
/linux-master/drivers/scsi/snic/ |
H A D | vnic_cq_fw.h | 21 desc = (struct snic_fw_req *)((u8 *)cq->ring.descs + 22 cq->ring.desc_size * cq->to_clean); 31 if (cq->to_clean == cq->ring.desc_count) { 36 desc = (struct snic_fw_req *)((u8 *)cq->ring.descs + 37 cq->ring.desc_size * cq->to_clean);
|
/linux-master/drivers/net/ethernet/synopsys/ |
H A D | dwc-xlgmac-desc.c | 75 struct xlgmac_ring *ring) 80 if (!ring) 83 if (ring->desc_data_head) { 84 for (i = 0; i < ring->dma_desc_count; i++) { 85 desc_data = XLGMAC_GET_DESC_DATA(ring, i); 89 kfree(ring->desc_data_head); 90 ring->desc_data_head = NULL; 93 if (ring->rx_hdr_pa.pages) { 94 dma_unmap_page(pdata->dev, ring->rx_hdr_pa.pages_dma, 95 ring 74 xlgmac_free_ring(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring) argument 125 xlgmac_init_ring(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring, unsigned int dma_desc_count) argument 389 xlgmac_map_rx_buffer(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring, struct xlgmac_desc_data *desc_data) argument 427 struct xlgmac_ring *ring; local 464 struct xlgmac_ring *ring; local 501 struct xlgmac_ring *ring = channel->tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_trace.h | 62 /* Events related to DIM, q_vectors and ring containers */ 117 /* Events related to a vsi & ring */ 119 TP_PROTO(struct ice_tx_ring *ring, struct ice_tx_desc *desc, 122 TP_ARGS(ring, desc, buf), 123 TP_STRUCT__entry(__field(void *, ring) 126 __string(devname, ring->netdev->name)), 128 TP_fast_assign(__entry->ring = ring; 131 __assign_str(devname, ring->netdev->name);), 133 TP_printk("netdev: %s ring [all...] |
H A D | ice_base.c | 248 * @vsi: VSI that ring belongs to 249 * @ring: ring to get the absolute queue index 252 static u16 ice_calc_txq_handle(struct ice_vsi *vsi, struct ice_tx_ring *ring, u8 tc) argument 254 WARN_ONCE(ice_ring_is_xdp(ring) && tc, "XDP ring can't belong to TC other than 0\n"); 256 if (ring->ch) 257 return ring->q_index - ring->ch->base_q; 260 * count from TC that ring belong 277 ice_eswitch_calc_txq_handle(struct ice_tx_ring *ring) argument 297 ice_cfg_xps_tx_ring(struct ice_tx_ring *ring) argument 319 ice_setup_tx_ctx(struct ice_tx_ring *ring, struct ice_tlan_ctx *tlan_ctx, u16 pf_q) argument 407 ice_setup_rx_ctx(struct ice_rx_ring *ring) argument 520 ice_xsk_pool_fill_cb(struct ice_rx_ring *ring) argument 539 ice_vsi_cfg_rxq(struct ice_rx_ring *ring) argument 886 ice_vsi_cfg_txq(struct ice_vsi *vsi, struct ice_tx_ring *ring, struct ice_aqc_add_tx_qgrp *qg_buf) argument 1132 ice_vsi_stop_tx_ring(struct ice_vsi *vsi, enum ice_disq_rst_src rst_src, u16 rel_vmvf_num, struct ice_tx_ring *ring, struct ice_txq_meta *txq_meta) argument 1190 ice_fill_txq_meta(const struct ice_vsi *vsi, struct ice_tx_ring *ring, struct ice_txq_meta *txq_meta) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | vcn_v2_0.c | 69 * Set ring and irq function pointers 97 struct amdgpu_ring *ring; local 128 ring = &adev->vcn.inst->ring_dec; 130 ring->use_doorbell = true; 131 ring->doorbell_index = adev->doorbell_index.vcn.vcn_ring0_1 << 1; 132 ring->vm_hub = AMDGPU_MMHUB0(0); 134 sprintf(ring->name, "vcn_dec"); 135 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst->irq, 0, 161 ring = &adev->vcn.inst->ring_enc[i]; 162 ring 229 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 798 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 933 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 1204 struct amdgpu_ring *ring; local 1324 vcn_v2_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 1338 vcn_v2_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 1355 vcn_v2_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument 1378 vcn_v2_0_dec_ring_insert_start(struct amdgpu_ring *ring) argument 1395 vcn_v2_0_dec_ring_insert_end(struct amdgpu_ring *ring) argument 1411 vcn_v2_0_dec_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 1434 vcn_v2_0_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 1473 vcn_v2_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 1492 vcn_v2_0_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1511 vcn_v2_0_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1526 vcn_v2_0_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1549 vcn_v2_0_enc_ring_get_rptr(struct amdgpu_ring *ring) argument 1566 vcn_v2_0_enc_ring_get_wptr(struct amdgpu_ring *ring) argument 1590 vcn_v2_0_enc_ring_set_wptr(struct amdgpu_ring *ring) argument 1621 vcn_v2_0_enc_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 1633 vcn_v2_0_enc_ring_insert_end(struct amdgpu_ring *ring) argument 1648 vcn_v2_0_enc_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 1662 vcn_v2_0_enc_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1671 vcn_v2_0_enc_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument 1684 vcn_v2_0_enc_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1724 vcn_v2_0_dec_ring_test_ring(struct amdgpu_ring *ring) argument 1857 struct amdgpu_ring *ring; local [all...] |
H A D | jpeg_v4_0_3.c | 45 static void jpeg_v4_0_3_dec_ring_set_wptr(struct amdgpu_ring *ring); 63 * Set ring and irq function pointers 88 struct amdgpu_ring *ring; local 111 ring = &adev->jpeg.inst[i].ring_dec[j]; 112 ring->use_doorbell = true; 113 ring->vm_hub = AMDGPU_MMHUB0(adev->jpeg.inst[i].aid_id); 115 ring->doorbell_index = 120 ring->doorbell_index = 124 ring->doorbell_index = 128 sprintf(ring 178 struct amdgpu_ring *ring; local 298 struct amdgpu_ring *ring; local 472 struct amdgpu_ring *ring; local 596 jpeg_v4_0_3_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 612 jpeg_v4_0_3_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 632 jpeg_v4_0_3_dec_ring_set_wptr(struct amdgpu_ring *ring) argument 655 jpeg_v4_0_3_dec_ring_insert_start(struct amdgpu_ring *ring) argument 675 jpeg_v4_0_3_dec_ring_insert_end(struct amdgpu_ring *ring) argument 698 jpeg_v4_0_3_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument 767 jpeg_v4_0_3_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 818 jpeg_v4_0_3_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 845 jpeg_v4_0_3_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument 860 jpeg_v4_0_3_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 878 jpeg_v4_0_3_dec_ring_nop(struct amdgpu_ring *ring, uint32_t count) argument [all...] |
H A D | sdma_v5_0.c | 252 static unsigned sdma_v5_0_ring_init_cond_exec(struct amdgpu_ring *ring, argument 257 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE)); 258 amdgpu_ring_write(ring, lower_32_bits(addr)); 259 amdgpu_ring_write(ring, upper_32_bits(addr)); 260 amdgpu_ring_write(ring, 1); 262 ret = ring->wptr & ring->buf_mask; 264 amdgpu_ring_write(ring, 0); 272 * @ring: amdgpu ring pointe 276 sdma_v5_0_ring_get_rptr(struct amdgpu_ring *ring) argument 294 sdma_v5_0_ring_get_wptr(struct amdgpu_ring *ring) argument 320 sdma_v5_0_ring_set_wptr(struct amdgpu_ring *ring) argument 387 sdma_v5_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 410 sdma_v5_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 445 sdma_v5_0_ring_emit_mem_sync(struct amdgpu_ring *ring) argument 469 sdma_v5_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 503 sdma_v5_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 668 struct amdgpu_ring *ring; local 969 sdma_v5_0_ring_test_ring(struct amdgpu_ring *ring) argument 1045 sdma_v5_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1226 sdma_v5_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1251 sdma_v5_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1280 sdma_v5_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1286 sdma_v5_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1295 sdma_v5_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1309 sdma_v5_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 1340 struct amdgpu_ring *ring; local 1474 sdma_v5_0_ring_preempt_ib(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_uvd.h | 45 struct amdgpu_ring ring; member in struct:amdgpu_uvd_inst 76 int amdgpu_uvd_entity_init(struct amdgpu_device *adev, struct amdgpu_ring *ring); 80 int amdgpu_uvd_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, 82 int amdgpu_uvd_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, 89 void amdgpu_uvd_ring_begin_use(struct amdgpu_ring *ring); 90 void amdgpu_uvd_ring_end_use(struct amdgpu_ring *ring); 91 int amdgpu_uvd_ring_test_ib(struct amdgpu_ring *ring, long timeout);
|
H A D | sdma_v5_2.c | 92 static unsigned sdma_v5_2_ring_init_cond_exec(struct amdgpu_ring *ring, argument 97 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE)); 98 amdgpu_ring_write(ring, lower_32_bits(addr)); 99 amdgpu_ring_write(ring, upper_32_bits(addr)); 100 amdgpu_ring_write(ring, 1); 102 ret = ring->wptr & ring->buf_mask; 104 amdgpu_ring_write(ring, 0); 112 * @ring: amdgpu ring pointe 116 sdma_v5_2_ring_get_rptr(struct amdgpu_ring *ring) argument 134 sdma_v5_2_ring_get_wptr(struct amdgpu_ring *ring) argument 160 sdma_v5_2_ring_set_wptr(struct amdgpu_ring *ring) argument 194 sdma_v5_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 217 sdma_v5_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 252 sdma_v5_2_ring_emit_mem_sync(struct amdgpu_ring *ring) argument 277 sdma_v5_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 312 sdma_v5_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 475 struct amdgpu_ring *ring; local 809 sdma_v5_2_ring_test_ring(struct amdgpu_ring *ring) argument 885 sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1066 sdma_v5_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1091 sdma_v5_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1120 sdma_v5_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1126 sdma_v5_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1135 sdma_v5_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1149 sdma_v5_2_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 1213 struct amdgpu_ring *ring; local 1331 sdma_v5_2_ring_preempt_ib(struct amdgpu_ring *ring) argument 1638 sdma_v5_2_ring_begin_use(struct amdgpu_ring *ring) argument 1653 sdma_v5_2_ring_end_use(struct amdgpu_ring *ring) argument [all...] |
H A D | sdma_v6_0.c | 83 static unsigned sdma_v6_0_ring_init_cond_exec(struct amdgpu_ring *ring, argument 88 amdgpu_ring_write(ring, SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_COND_EXE)); 89 amdgpu_ring_write(ring, lower_32_bits(addr)); 90 amdgpu_ring_write(ring, upper_32_bits(addr)); 91 amdgpu_ring_write(ring, 1); 93 ret = ring->wptr & ring->buf_mask; 95 amdgpu_ring_write(ring, 0); 103 * @ring: amdgpu ring pointe 107 sdma_v6_0_ring_get_rptr(struct amdgpu_ring *ring) argument 125 sdma_v6_0_ring_get_wptr(struct amdgpu_ring *ring) argument 145 sdma_v6_0_ring_set_wptr(struct amdgpu_ring *ring) argument 180 sdma_v6_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 203 sdma_v6_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 238 sdma_v6_0_ring_emit_mem_sync(struct amdgpu_ring *ring) argument 263 sdma_v6_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 294 sdma_v6_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 423 struct amdgpu_ring *ring; local 710 struct amdgpu_ring *ring; local 824 sdma_v6_0_ring_test_ring(struct amdgpu_ring *ring) argument 900 sdma_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 1080 sdma_v6_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 1104 sdma_v6_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 1132 sdma_v6_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1159 sdma_v6_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1168 sdma_v6_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1182 sdma_v6_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument 1230 struct amdgpu_ring *ring; local 1350 sdma_v6_0_ring_preempt_ib(struct amdgpu_ring *ring) argument [all...] |
H A D | uvd_v5_0.c | 52 * @ring: amdgpu_ring pointer 56 static uint64_t uvd_v5_0_ring_get_rptr(struct amdgpu_ring *ring) argument 58 struct amdgpu_device *adev = ring->adev; 66 * @ring: amdgpu_ring pointer 70 static uint64_t uvd_v5_0_ring_get_wptr(struct amdgpu_ring *ring) argument 72 struct amdgpu_device *adev = ring->adev; 80 * @ring: amdgpu_ring pointer 84 static void uvd_v5_0_ring_set_wptr(struct amdgpu_ring *ring) argument 86 struct amdgpu_device *adev = ring->adev; 88 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring 104 struct amdgpu_ring *ring; local 153 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local 322 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local 493 uvd_v5_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 522 uvd_v5_0_ring_test_ring(struct amdgpu_ring *ring) argument 559 uvd_v5_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 572 uvd_v5_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument [all...] |
H A D | sdma_v2_4.c | 82 * and each one supports 1 ring buffer used for gfx 86 * (ring buffer, IBs, etc.), but sDMA has it's own 182 * @ring: amdgpu ring pointer 186 static uint64_t sdma_v2_4_ring_get_rptr(struct amdgpu_ring *ring) argument 189 return *ring->rptr_cpu_addr >> 2; 195 * @ring: amdgpu ring pointer 199 static uint64_t sdma_v2_4_ring_get_wptr(struct amdgpu_ring *ring) argument 201 struct amdgpu_device *adev = ring 214 sdma_v2_4_ring_set_wptr(struct amdgpu_ring *ring) argument 221 sdma_v2_4_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 244 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 272 sdma_v2_4_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 304 sdma_v2_4_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 400 struct amdgpu_ring *ring; local 528 sdma_v2_4_ring_test_ring(struct amdgpu_ring *ring) argument 581 sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 731 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 755 sdma_v2_4_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 783 sdma_v2_4_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 800 sdma_v2_4_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 830 struct amdgpu_ring *ring; local [all...] |
H A D | vcn_v1_0.c | 57 static void vcn_v1_0_ring_begin_use(struct amdgpu_ring *ring); 64 * Set ring and irq function pointers 91 struct amdgpu_ring *ring; local 122 ring = &adev->vcn.inst->ring_dec; 123 ring->vm_hub = AMDGPU_MMHUB0(0); 124 sprintf(ring->name, "vcn_dec"); 125 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst->irq, 0, 144 ring = &adev->vcn.inst->ring_enc[i]; 145 ring->vm_hub = AMDGPU_MMHUB0(0); 146 sprintf(ring 200 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 791 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 965 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local 1217 struct amdgpu_ring *ring; local 1382 vcn_v1_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 1396 vcn_v1_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 1410 vcn_v1_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument 1428 vcn_v1_0_dec_ring_insert_start(struct amdgpu_ring *ring) argument 1447 vcn_v1_0_dec_ring_insert_end(struct amdgpu_ring *ring) argument 1466 vcn_v1_0_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 1507 vcn_v1_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 1530 vcn_v1_0_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1550 vcn_v1_0_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 1565 vcn_v1_0_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1588 vcn_v1_0_enc_ring_get_rptr(struct amdgpu_ring *ring) argument 1605 vcn_v1_0_enc_ring_get_wptr(struct amdgpu_ring *ring) argument 1622 vcn_v1_0_enc_ring_set_wptr(struct amdgpu_ring *ring) argument 1644 vcn_v1_0_enc_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 1656 vcn_v1_0_enc_ring_insert_end(struct amdgpu_ring *ring) argument 1671 vcn_v1_0_enc_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 1685 vcn_v1_0_enc_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument 1695 vcn_v1_0_enc_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument 1708 vcn_v1_0_enc_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 1749 vcn_v1_0_dec_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 1828 vcn_v1_0_ring_begin_use(struct amdgpu_ring *ring) argument 1842 vcn_v1_0_set_pg_for_begin_use(struct amdgpu_ring *ring, bool set_clocks) argument 1881 vcn_v1_0_ring_end_use(struct amdgpu_ring *ring) argument [all...] |
H A D | uvd_v3_1.c | 40 * @ring: amdgpu_ring pointer 44 static uint64_t uvd_v3_1_ring_get_rptr(struct amdgpu_ring *ring) argument 46 struct amdgpu_device *adev = ring->adev; 54 * @ring: amdgpu_ring pointer 58 static uint64_t uvd_v3_1_ring_get_wptr(struct amdgpu_ring *ring) argument 60 struct amdgpu_device *adev = ring->adev; 68 * @ring: amdgpu_ring pointer 72 static void uvd_v3_1_ring_set_wptr(struct amdgpu_ring *ring) argument 74 struct amdgpu_device *adev = ring->adev; 76 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring 89 uvd_v3_1_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 110 uvd_v3_1_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 139 uvd_v3_1_ring_test_ring(struct amdgpu_ring *ring) argument 167 uvd_v3_1_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 322 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local 547 struct amdgpu_ring *ring; local 631 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local [all...] |
H A D | uvd_v4_2.c | 54 * @ring: amdgpu_ring pointer 58 static uint64_t uvd_v4_2_ring_get_rptr(struct amdgpu_ring *ring) argument 60 struct amdgpu_device *adev = ring->adev; 68 * @ring: amdgpu_ring pointer 72 static uint64_t uvd_v4_2_ring_get_wptr(struct amdgpu_ring *ring) argument 74 struct amdgpu_device *adev = ring->adev; 82 * @ring: amdgpu_ring pointer 86 static void uvd_v4_2_ring_set_wptr(struct amdgpu_ring *ring) argument 88 struct amdgpu_device *adev = ring->adev; 90 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring 106 struct amdgpu_ring *ring; local 157 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local 285 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local 477 uvd_v4_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 506 uvd_v4_2_ring_test_ring(struct amdgpu_ring *ring) argument 544 uvd_v4_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 555 uvd_v4_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument [all...] |
H A D | cik_sdma.c | 86 * and each one supports 1 ring buffer used for gfx 90 * (ring buffer, IBs, etc.), but sDMA has it's own 155 * @ring: amdgpu ring pointer 159 static uint64_t cik_sdma_ring_get_rptr(struct amdgpu_ring *ring) argument 163 rptr = *ring->rptr_cpu_addr; 171 * @ring: amdgpu ring pointer 175 static uint64_t cik_sdma_ring_get_wptr(struct amdgpu_ring *ring) argument 177 struct amdgpu_device *adev = ring 189 cik_sdma_ring_set_wptr(struct amdgpu_ring *ring) argument 197 cik_sdma_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument 220 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument 245 cik_sdma_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument 276 cik_sdma_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument 426 struct amdgpu_ring *ring; local 598 cik_sdma_ring_test_ring(struct amdgpu_ring *ring) argument 650 cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument 796 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument 820 cik_sdma_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument 847 cik_sdma_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument 863 cik_sdma_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument 941 struct amdgpu_ring *ring; local [all...] |
/linux-master/drivers/net/can/spi/mcp251xfd/ |
H A D | mcp251xfd-chip-fifo.c | 21 const struct mcp251xfd_rx_ring *ring) 32 ring->obj_num - 1) | 45 MCP251XFD_REG_FIFOCON(ring->fifo_nr), fifo_con); 50 const struct mcp251xfd_rx_ring *ring) 54 fltcon = MCP251XFD_REG_FLTCON_FLTEN(ring->nr) | 55 MCP251XFD_REG_FLTCON_FBP(ring->nr, ring->fifo_nr); 58 MCP251XFD_REG_FLTCON(ring->nr >> 2), 59 MCP251XFD_REG_FLTCON_FLT_MASK(ring->nr), 20 mcp251xfd_chip_rx_fifo_init_one(const struct mcp251xfd_priv *priv, const struct mcp251xfd_rx_ring *ring) argument 49 mcp251xfd_chip_rx_filter_init_one(const struct mcp251xfd_priv *priv, const struct mcp251xfd_rx_ring *ring) argument
|
/linux-master/drivers/mailbox/ |
H A D | bcm-flexrm-mailbox.c | 8 * extension to Broadcom FlexRM ring manager. The FlexRM ring 13 * rings where each mailbox channel represents a separate FlexRM ring. 128 /* ====== FlexRM ring descriptor defines ===== */ 292 /* ====== FlexRM ring descriptor helper routines ===== */ 924 struct flexrm_ring *ring; local 931 ring = &mbox->rings[i]; 932 if (readl(ring->regs + RING_CONTROL) & 939 ring->num, state, 940 (unsigned long long)ring 952 struct flexrm_ring *ring; local 975 flexrm_new_request(struct flexrm_ring *ring, struct brcm_message *batch_msg, struct brcm_message *msg) argument 1074 flexrm_process_completions(struct flexrm_ring *ring) argument 1196 struct flexrm_ring *ring = chan->con_priv; local 1229 struct flexrm_ring *ring = chan->con_priv; local 1364 struct flexrm_ring *ring = chan->con_priv; local 1449 struct flexrm_ring *ring; local 1477 struct flexrm_ring *ring = &mbox->rings[desc->msi_index]; local 1491 struct flexrm_ring *ring; local [all...] |
/linux-master/net/9p/ |
H A D | trans_xen.c | 24 #define XEN_9PFS_RING_SIZE(ring) XEN_FLEX_RING_SIZE(ring->intf->ring_order) 34 /* One per ring, more than one per 9pfs share */ 42 /* protect a ring from concurrent accesses */ 103 static bool p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size) argument 107 cons = ring->intf->out_cons; 108 prod = ring->intf->out_prod; 111 return XEN_9PFS_RING_SIZE(ring) - 112 xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE(ring)) >= size; 121 struct xen_9pfs_dataring *ring; local 173 struct xen_9pfs_dataring *ring; local 243 struct xen_9pfs_dataring *ring = r; local 282 struct xen_9pfs_dataring *ring = &priv->rings[i]; local 319 xen_9pfs_front_alloc_dataring(struct xenbus_device *dev, struct xen_9pfs_dataring *ring, unsigned int order) argument [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | vce_v1_0.c | 56 * @ring: radeon_ring pointer 61 struct radeon_ring *ring) 63 if (ring->idx == TN_RING_TYPE_VCE1_INDEX) 73 * @ring: radeon_ring pointer 78 struct radeon_ring *ring) 80 if (ring->idx == TN_RING_TYPE_VCE1_INDEX) 90 * @ring: radeon_ring pointer 95 struct radeon_ring *ring) 97 if (ring->idx == TN_RING_TYPE_VCE1_INDEX) 98 WREG32(VCE_RB_WPTR, ring 60 vce_v1_0_get_rptr(struct radeon_device *rdev, struct radeon_ring *ring) argument 77 vce_v1_0_get_wptr(struct radeon_device *rdev, struct radeon_ring *ring) argument 94 vce_v1_0_set_wptr(struct radeon_device *rdev, struct radeon_ring *ring) argument 291 struct radeon_ring *ring; local 358 struct radeon_ring *ring; local [all...] |
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a5xx_preempt.c | 39 /* Write the most recent wptr for the given ring into the hardware */ 40 static inline void update_wptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument 45 if (!ring) 48 spin_lock_irqsave(&ring->preempt_lock, flags); 49 wptr = get_wptr(ring); 50 spin_unlock_irqrestore(&ring->preempt_lock, flags); 63 struct msm_ringbuffer *ring = gpu->rb[i]; local 65 spin_lock_irqsave(&ring->preempt_lock, flags); 66 empty = (get_wptr(ring) == gpu->funcs->get_rptr(gpu, ring)); 95 struct msm_ringbuffer *ring; local 220 preempt_init_ring(struct a5xx_gpu *a5xx_gpu, struct msm_ringbuffer *ring) argument [all...] |