Searched refs:ring (Results 76 - 100 of 804) sorted by relevance

1234567891011>>

/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c75 struct mlx4_en_rx_ring *ring,
86 ring->rx_alloc_pages++;
109 struct mlx4_en_rx_ring *ring, int index)
111 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index;
122 /* If the number of used fragments does not fill up the ring stride,
125 possible_frags = (ring->stride - sizeof(struct mlx4_en_rx_desc)) / DS_SIZE;
134 struct mlx4_en_rx_ring *ring, int index,
137 struct mlx4_en_rx_desc *rx_desc = ring->buf +
138 (index << ring
74 mlx4_en_alloc_frags(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, gfp_t gfp) argument
108 mlx4_en_init_rx_desc(const struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument
133 mlx4_en_prepare_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index, gfp_t gfp) argument
157 mlx4_en_is_ring_empty(const struct mlx4_en_rx_ring *ring) argument
162 mlx4_en_update_rx_prod_db(struct mlx4_en_rx_ring *ring) argument
168 mlx4_en_free_rx_desc(const struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument
185 struct mlx4_en_rx_ring *ring; local
226 mlx4_en_free_rx_buf(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
268 struct mlx4_en_rx_ring *ring; local
328 struct mlx4_en_rx_ring *ring; local
395 int ring; local
415 mlx4_en_rx_recycle(struct mlx4_en_rx_ring *ring, struct mlx4_en_rx_alloc *frame) argument
434 struct mlx4_en_rx_ring *ring = *pring; local
450 mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
547 mlx4_en_refill_rx_buffers(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
668 struct mlx4_en_rx_ring *ring; member in struct:mlx4_en_xdp_buff
718 struct mlx4_en_rx_ring *ring; local
1114 mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, int qpn, struct mlx4_en_rx_ring *ring, enum mlx4_qp_state *state, struct mlx4_qp *qp) argument
[all...]
/linux-master/drivers/net/ethernet/hisilicon/hns3/
H A Dhns3_enet.c428 if (tqp_vectors->tx_group.ring && tqp_vectors->rx_group.ring) {
434 } else if (tqp_vectors->rx_group.ring) {
439 } else if (tqp_vectors->tx_group.ring) {
751 struct hns3_enet_ring *ring = tqp_vector->tx_group.ring; local
753 while (ring) {
758 ring->tqp->tqp_index);
763 ring = ring
962 hns3_tx_spare_space(struct hns3_enet_ring *ring) argument
982 hns3_tx_spare_update(struct hns3_enet_ring *ring) argument
997 hns3_can_use_tx_bounce(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 space) argument
1015 hns3_can_use_tx_sgl(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 space) argument
1032 hns3_init_tx_spare_buffer(struct hns3_enet_ring *ring) argument
1089 hns3_tx_spare_alloc(struct hns3_enet_ring *ring, unsigned int size, dma_addr_t *dma, u32 *cb_len) argument
1116 hns3_tx_spare_rollback(struct hns3_enet_ring *ring, u32 len) argument
1128 hns3_tx_spare_reclaim_cb(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument
1579 hns3_handle_vlan_info(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_param *param) argument
1604 hns3_handle_csum_partial(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_cb *desc_cb, struct hns3_desc_param *param) argument
1648 hns3_fill_skb_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc *desc, struct hns3_desc_cb *desc_cb) argument
1680 hns3_fill_desc(struct hns3_enet_ring *ring, dma_addr_t dma, unsigned int size) argument
1723 hns3_map_and_fill_desc(struct hns3_enet_ring *ring, void *priv, unsigned int type) argument
1893 hns3_skb_linearize(struct hns3_enet_ring *ring, struct sk_buff *skb, unsigned int bd_num) argument
1922 hns3_nic_maybe_stop_tx(struct hns3_enet_ring *ring, struct net_device *netdev, struct sk_buff *skb) argument
1970 hns3_clear_desc(struct hns3_enet_ring *ring, int next_to_use_orig) argument
2010 hns3_fill_skb_to_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, unsigned int type) argument
2044 hns3_tx_push_bd(struct hns3_enet_ring *ring, int num) argument
2075 hns3_tx_mem_doorbell(struct hns3_enet_ring *ring) argument
2095 hns3_tx_doorbell(struct hns3_enet_ring *ring, int num, bool doorbell) argument
2148 hns3_handle_tx_bounce(struct hns3_enet_ring *ring, struct sk_buff *skb) argument
2201 hns3_handle_tx_sgl(struct hns3_enet_ring *ring, struct sk_buff *skb) argument
2253 hns3_handle_desc_filling(struct hns3_enet_ring *ring, struct sk_buff *skb) argument
2273 hns3_handle_skb_desc(struct hns3_enet_ring *ring, struct sk_buff *skb, struct hns3_desc_cb *desc_cb, int next_to_use_head) argument
2301 struct hns3_enet_ring *ring = &priv->ring[skb->queue_mapping]; local
2494 hns3_fetch_stats(struct rtnl_link_stats64 *stats, struct hns3_enet_ring *ring, bool is_tx) argument
2544 struct hns3_enet_ring *ring; local
3367 hns3_alloc_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument
3404 hns3_free_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb, int budget) argument
3420 hns3_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument
3431 hns3_unmap_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument
3445 hns3_buffer_detach(struct hns3_enet_ring *ring, int i) argument
3452 hns3_free_buffer_detach(struct hns3_enet_ring *ring, int i, int budget) argument
3464 hns3_free_buffers(struct hns3_enet_ring *ring) argument
3473 hns3_free_desc(struct hns3_enet_ring *ring) argument
3486 hns3_alloc_desc(struct hns3_enet_ring *ring) argument
3498 hns3_alloc_and_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) argument
3519 hns3_alloc_and_attach_buffer(struct hns3_enet_ring *ring, int i) argument
3534 hns3_alloc_ring_buffers(struct hns3_enet_ring *ring) argument
3553 hns3_replace_buffer(struct hns3_enet_ring *ring, int i, struct hns3_desc_cb *res_cb) argument
3564 hns3_reuse_buffer(struct hns3_enet_ring *ring, int i) argument
3578 hns3_nic_reclaim_desc(struct hns3_enet_ring *ring, int *bytes, int *pkts, int budget) argument
3630 hns3_clean_tx_ring(struct hns3_enet_ring *ring, int budget) argument
3668 hns3_desc_unused(struct hns3_enet_ring *ring) argument
3680 hns3_nic_alloc_rx_buffers(struct hns3_enet_ring *ring, int cleand_count) argument
3723 hns3_handle_rx_copybreak(struct sk_buff *skb, int i, struct hns3_enet_ring *ring, int pull_len, struct hns3_desc_cb *desc_cb) argument
3751 hns3_nic_reuse_page(struct sk_buff *skb, int i, struct hns3_enet_ring *ring, int pull_len, struct hns3_desc_cb *desc_cb) argument
3873 hns3_checksum_complete(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 ptype, u16 csum) argument
3923 hns3_rx_checksum(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 l234info, u32 bd_base_info, u32 ol_info, u16 csum) argument
3960 hns3_rx_skb(struct hns3_enet_ring *ring, struct sk_buff *skb) argument
3968 hns3_parse_vlan_tag(struct hns3_enet_ring *ring, struct hns3_desc *desc, u32 l234info, u16 *vlan_tag) argument
4021 hns3_rx_ring_move_fw(struct hns3_enet_ring *ring) argument
4032 hns3_alloc_skb(struct hns3_enet_ring *ring, unsigned int length, unsigned char *va) argument
4085 hns3_add_frag(struct hns3_enet_ring *ring) argument
4146 hns3_set_gro_and_checksum(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 l234info, u32 bd_base_info, u32 ol_info, u16 csum) argument
4188 hns3_set_rx_skb_rss_type(struct hns3_enet_ring *ring, struct sk_buff *skb, u32 rss_hash, u32 l234info, u32 ol_info) argument
4236 hns3_handle_rx_vlan_tag(struct hns3_enet_ring *ring, struct hns3_desc *desc, struct sk_buff *skb, u32 l234info) argument
4255 hns3_handle_bdinfo(struct hns3_enet_ring *ring, struct sk_buff *skb) argument
4325 hns3_handle_rx_bd(struct hns3_enet_ring *ring) argument
4397 hns3_clean_rx_ring(struct hns3_enet_ring *ring, int budget, void (*rx_fn)(struct hns3_enet_ring *, struct sk_buff *)) argument
4471 struct hns3_enet_ring *ring; local
4528 struct hns3_enet_ring *ring; local
4600 hns3_add_ring_to_group(struct hns3_enet_ring_group *group, struct hns3_enet_ring *ring) argument
4876 struct hns3_enet_ring *ring; local
4940 hns3_alloc_page_pool(struct hns3_enet_ring *ring) argument
4962 hns3_alloc_ring_memory(struct hns3_enet_ring *ring) argument
5002 hns3_fini_ring(struct hns3_enet_ring *ring) argument
5055 hns3_init_ring_hw(struct hns3_enet_ring *ring) argument
5482 hns3_clear_tx_ring(struct hns3_enet_ring *ring) argument
5493 hns3_clear_rx_ring(struct hns3_enet_ring *ring) argument
5530 hns3_force_clear_rx_ring(struct hns3_enet_ring *ring) argument
5554 struct hns3_enet_ring *ring; local
[all...]
/linux-master/drivers/scsi/fnic/
H A Dvnic_cq_copy.h23 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs +
24 cq->ring.desc_size * cq->to_clean);
33 if (cq->to_clean == cq->ring.desc_count) {
38 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs +
39 cq->ring.desc_size * cq->to_clean);
H A Dvnic_wq_copy.h19 struct vnic_dev_ring ring; member in struct:vnic_wq_copy
26 return wq->ring.desc_avail;
31 return wq->ring.desc_count - 1 - wq->ring.desc_avail;
36 struct fcpio_host_req *desc = wq->ring.descs;
43 ((wq->to_use_index + 1) == wq->ring.desc_count) ?
45 wq->ring.desc_avail--;
64 cnt = wq->ring.desc_count - wq->to_clean_index + index + 1;
66 wq->to_clean_index = ((index + 1) % wq->ring.desc_count);
67 wq->ring
[all...]
/linux-master/drivers/scsi/snic/
H A Dvnic_cq_fw.h21 desc = (struct snic_fw_req *)((u8 *)cq->ring.descs +
22 cq->ring.desc_size * cq->to_clean);
31 if (cq->to_clean == cq->ring.desc_count) {
36 desc = (struct snic_fw_req *)((u8 *)cq->ring.descs +
37 cq->ring.desc_size * cq->to_clean);
/linux-master/drivers/net/ethernet/synopsys/
H A Ddwc-xlgmac-desc.c75 struct xlgmac_ring *ring)
80 if (!ring)
83 if (ring->desc_data_head) {
84 for (i = 0; i < ring->dma_desc_count; i++) {
85 desc_data = XLGMAC_GET_DESC_DATA(ring, i);
89 kfree(ring->desc_data_head);
90 ring->desc_data_head = NULL;
93 if (ring->rx_hdr_pa.pages) {
94 dma_unmap_page(pdata->dev, ring->rx_hdr_pa.pages_dma,
95 ring
74 xlgmac_free_ring(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring) argument
125 xlgmac_init_ring(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring, unsigned int dma_desc_count) argument
389 xlgmac_map_rx_buffer(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring, struct xlgmac_desc_data *desc_data) argument
427 struct xlgmac_ring *ring; local
464 struct xlgmac_ring *ring; local
501 struct xlgmac_ring *ring = channel->tx_ring; local
[all...]
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_trace.h62 /* Events related to DIM, q_vectors and ring containers */
117 /* Events related to a vsi & ring */
119 TP_PROTO(struct ice_tx_ring *ring, struct ice_tx_desc *desc,
122 TP_ARGS(ring, desc, buf),
123 TP_STRUCT__entry(__field(void *, ring)
126 __string(devname, ring->netdev->name)),
128 TP_fast_assign(__entry->ring = ring;
131 __assign_str(devname, ring->netdev->name);),
133 TP_printk("netdev: %s ring
[all...]
H A Dice_base.c248 * @vsi: VSI that ring belongs to
249 * @ring: ring to get the absolute queue index
252 static u16 ice_calc_txq_handle(struct ice_vsi *vsi, struct ice_tx_ring *ring, u8 tc) argument
254 WARN_ONCE(ice_ring_is_xdp(ring) && tc, "XDP ring can't belong to TC other than 0\n");
256 if (ring->ch)
257 return ring->q_index - ring->ch->base_q;
260 * count from TC that ring belong
277 ice_eswitch_calc_txq_handle(struct ice_tx_ring *ring) argument
297 ice_cfg_xps_tx_ring(struct ice_tx_ring *ring) argument
319 ice_setup_tx_ctx(struct ice_tx_ring *ring, struct ice_tlan_ctx *tlan_ctx, u16 pf_q) argument
407 ice_setup_rx_ctx(struct ice_rx_ring *ring) argument
520 ice_xsk_pool_fill_cb(struct ice_rx_ring *ring) argument
539 ice_vsi_cfg_rxq(struct ice_rx_ring *ring) argument
886 ice_vsi_cfg_txq(struct ice_vsi *vsi, struct ice_tx_ring *ring, struct ice_aqc_add_tx_qgrp *qg_buf) argument
1132 ice_vsi_stop_tx_ring(struct ice_vsi *vsi, enum ice_disq_rst_src rst_src, u16 rel_vmvf_num, struct ice_tx_ring *ring, struct ice_txq_meta *txq_meta) argument
1190 ice_fill_txq_meta(const struct ice_vsi *vsi, struct ice_tx_ring *ring, struct ice_txq_meta *txq_meta) argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Dvcn_v2_0.c69 * Set ring and irq function pointers
97 struct amdgpu_ring *ring; local
128 ring = &adev->vcn.inst->ring_dec;
130 ring->use_doorbell = true;
131 ring->doorbell_index = adev->doorbell_index.vcn.vcn_ring0_1 << 1;
132 ring->vm_hub = AMDGPU_MMHUB0(0);
134 sprintf(ring->name, "vcn_dec");
135 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst->irq, 0,
161 ring = &adev->vcn.inst->ring_enc[i];
162 ring
229 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
798 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
933 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
1204 struct amdgpu_ring *ring; local
1324 vcn_v2_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
1338 vcn_v2_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
1355 vcn_v2_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
1378 vcn_v2_0_dec_ring_insert_start(struct amdgpu_ring *ring) argument
1395 vcn_v2_0_dec_ring_insert_end(struct amdgpu_ring *ring) argument
1411 vcn_v2_0_dec_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
1434 vcn_v2_0_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
1473 vcn_v2_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
1492 vcn_v2_0_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1511 vcn_v2_0_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1526 vcn_v2_0_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1549 vcn_v2_0_enc_ring_get_rptr(struct amdgpu_ring *ring) argument
1566 vcn_v2_0_enc_ring_get_wptr(struct amdgpu_ring *ring) argument
1590 vcn_v2_0_enc_ring_set_wptr(struct amdgpu_ring *ring) argument
1621 vcn_v2_0_enc_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
1633 vcn_v2_0_enc_ring_insert_end(struct amdgpu_ring *ring) argument
1648 vcn_v2_0_enc_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
1662 vcn_v2_0_enc_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1671 vcn_v2_0_enc_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument
1684 vcn_v2_0_enc_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1724 vcn_v2_0_dec_ring_test_ring(struct amdgpu_ring *ring) argument
1857 struct amdgpu_ring *ring; local
[all...]
H A Djpeg_v4_0_3.c45 static void jpeg_v4_0_3_dec_ring_set_wptr(struct amdgpu_ring *ring);
63 * Set ring and irq function pointers
88 struct amdgpu_ring *ring; local
111 ring = &adev->jpeg.inst[i].ring_dec[j];
112 ring->use_doorbell = true;
113 ring->vm_hub = AMDGPU_MMHUB0(adev->jpeg.inst[i].aid_id);
115 ring->doorbell_index =
120 ring->doorbell_index =
124 ring->doorbell_index =
128 sprintf(ring
178 struct amdgpu_ring *ring; local
298 struct amdgpu_ring *ring; local
472 struct amdgpu_ring *ring; local
596 jpeg_v4_0_3_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
612 jpeg_v4_0_3_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
632 jpeg_v4_0_3_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
655 jpeg_v4_0_3_dec_ring_insert_start(struct amdgpu_ring *ring) argument
675 jpeg_v4_0_3_dec_ring_insert_end(struct amdgpu_ring *ring) argument
698 jpeg_v4_0_3_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned int flags) argument
767 jpeg_v4_0_3_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
818 jpeg_v4_0_3_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
845 jpeg_v4_0_3_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument
860 jpeg_v4_0_3_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
878 jpeg_v4_0_3_dec_ring_nop(struct amdgpu_ring *ring, uint32_t count) argument
[all...]
H A Dsdma_v5_0.c252 static unsigned sdma_v5_0_ring_init_cond_exec(struct amdgpu_ring *ring, argument
257 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE));
258 amdgpu_ring_write(ring, lower_32_bits(addr));
259 amdgpu_ring_write(ring, upper_32_bits(addr));
260 amdgpu_ring_write(ring, 1);
262 ret = ring->wptr & ring->buf_mask;
264 amdgpu_ring_write(ring, 0);
272 * @ring: amdgpu ring pointe
276 sdma_v5_0_ring_get_rptr(struct amdgpu_ring *ring) argument
294 sdma_v5_0_ring_get_wptr(struct amdgpu_ring *ring) argument
320 sdma_v5_0_ring_set_wptr(struct amdgpu_ring *ring) argument
387 sdma_v5_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
410 sdma_v5_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
445 sdma_v5_0_ring_emit_mem_sync(struct amdgpu_ring *ring) argument
469 sdma_v5_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
503 sdma_v5_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
668 struct amdgpu_ring *ring; local
969 sdma_v5_0_ring_test_ring(struct amdgpu_ring *ring) argument
1045 sdma_v5_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1226 sdma_v5_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1251 sdma_v5_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1280 sdma_v5_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1286 sdma_v5_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1295 sdma_v5_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1309 sdma_v5_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
1340 struct amdgpu_ring *ring; local
1474 sdma_v5_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
[all...]
H A Damdgpu_uvd.h45 struct amdgpu_ring ring; member in struct:amdgpu_uvd_inst
76 int amdgpu_uvd_entity_init(struct amdgpu_device *adev, struct amdgpu_ring *ring);
80 int amdgpu_uvd_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
82 int amdgpu_uvd_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
89 void amdgpu_uvd_ring_begin_use(struct amdgpu_ring *ring);
90 void amdgpu_uvd_ring_end_use(struct amdgpu_ring *ring);
91 int amdgpu_uvd_ring_test_ib(struct amdgpu_ring *ring, long timeout);
H A Dsdma_v5_2.c92 static unsigned sdma_v5_2_ring_init_cond_exec(struct amdgpu_ring *ring, argument
97 amdgpu_ring_write(ring, SDMA_PKT_HEADER_OP(SDMA_OP_COND_EXE));
98 amdgpu_ring_write(ring, lower_32_bits(addr));
99 amdgpu_ring_write(ring, upper_32_bits(addr));
100 amdgpu_ring_write(ring, 1);
102 ret = ring->wptr & ring->buf_mask;
104 amdgpu_ring_write(ring, 0);
112 * @ring: amdgpu ring pointe
116 sdma_v5_2_ring_get_rptr(struct amdgpu_ring *ring) argument
134 sdma_v5_2_ring_get_wptr(struct amdgpu_ring *ring) argument
160 sdma_v5_2_ring_set_wptr(struct amdgpu_ring *ring) argument
194 sdma_v5_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
217 sdma_v5_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
252 sdma_v5_2_ring_emit_mem_sync(struct amdgpu_ring *ring) argument
277 sdma_v5_2_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
312 sdma_v5_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
475 struct amdgpu_ring *ring; local
809 sdma_v5_2_ring_test_ring(struct amdgpu_ring *ring) argument
885 sdma_v5_2_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1066 sdma_v5_2_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1091 sdma_v5_2_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1120 sdma_v5_2_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1126 sdma_v5_2_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1135 sdma_v5_2_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1149 sdma_v5_2_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
1213 struct amdgpu_ring *ring; local
1331 sdma_v5_2_ring_preempt_ib(struct amdgpu_ring *ring) argument
1638 sdma_v5_2_ring_begin_use(struct amdgpu_ring *ring) argument
1653 sdma_v5_2_ring_end_use(struct amdgpu_ring *ring) argument
[all...]
H A Dsdma_v6_0.c83 static unsigned sdma_v6_0_ring_init_cond_exec(struct amdgpu_ring *ring, argument
88 amdgpu_ring_write(ring, SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_COND_EXE));
89 amdgpu_ring_write(ring, lower_32_bits(addr));
90 amdgpu_ring_write(ring, upper_32_bits(addr));
91 amdgpu_ring_write(ring, 1);
93 ret = ring->wptr & ring->buf_mask;
95 amdgpu_ring_write(ring, 0);
103 * @ring: amdgpu ring pointe
107 sdma_v6_0_ring_get_rptr(struct amdgpu_ring *ring) argument
125 sdma_v6_0_ring_get_wptr(struct amdgpu_ring *ring) argument
145 sdma_v6_0_ring_set_wptr(struct amdgpu_ring *ring) argument
180 sdma_v6_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
203 sdma_v6_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
238 sdma_v6_0_ring_emit_mem_sync(struct amdgpu_ring *ring) argument
263 sdma_v6_0_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
294 sdma_v6_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
423 struct amdgpu_ring *ring; local
710 struct amdgpu_ring *ring; local
824 sdma_v6_0_ring_test_ring(struct amdgpu_ring *ring) argument
900 sdma_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
1080 sdma_v6_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
1104 sdma_v6_0_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
1132 sdma_v6_0_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1159 sdma_v6_0_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1168 sdma_v6_0_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1182 sdma_v6_0_ring_emit_reg_write_reg_wait(struct amdgpu_ring *ring, uint32_t reg0, uint32_t reg1, uint32_t ref, uint32_t mask) argument
1230 struct amdgpu_ring *ring; local
1350 sdma_v6_0_ring_preempt_ib(struct amdgpu_ring *ring) argument
[all...]
H A Duvd_v5_0.c52 * @ring: amdgpu_ring pointer
56 static uint64_t uvd_v5_0_ring_get_rptr(struct amdgpu_ring *ring) argument
58 struct amdgpu_device *adev = ring->adev;
66 * @ring: amdgpu_ring pointer
70 static uint64_t uvd_v5_0_ring_get_wptr(struct amdgpu_ring *ring) argument
72 struct amdgpu_device *adev = ring->adev;
80 * @ring: amdgpu_ring pointer
84 static void uvd_v5_0_ring_set_wptr(struct amdgpu_ring *ring) argument
86 struct amdgpu_device *adev = ring->adev;
88 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring
104 struct amdgpu_ring *ring; local
153 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
322 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
493 uvd_v5_0_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
522 uvd_v5_0_ring_test_ring(struct amdgpu_ring *ring) argument
559 uvd_v5_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
572 uvd_v5_0_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
[all...]
H A Dsdma_v2_4.c82 * and each one supports 1 ring buffer used for gfx
86 * (ring buffer, IBs, etc.), but sDMA has it's own
182 * @ring: amdgpu ring pointer
186 static uint64_t sdma_v2_4_ring_get_rptr(struct amdgpu_ring *ring) argument
189 return *ring->rptr_cpu_addr >> 2;
195 * @ring: amdgpu ring pointer
199 static uint64_t sdma_v2_4_ring_get_wptr(struct amdgpu_ring *ring) argument
201 struct amdgpu_device *adev = ring
214 sdma_v2_4_ring_set_wptr(struct amdgpu_ring *ring) argument
221 sdma_v2_4_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
244 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
272 sdma_v2_4_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
304 sdma_v2_4_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
400 struct amdgpu_ring *ring; local
528 sdma_v2_4_ring_test_ring(struct amdgpu_ring *ring) argument
581 sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
731 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
755 sdma_v2_4_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
783 sdma_v2_4_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
800 sdma_v2_4_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
830 struct amdgpu_ring *ring; local
[all...]
H A Dvcn_v1_0.c57 static void vcn_v1_0_ring_begin_use(struct amdgpu_ring *ring);
64 * Set ring and irq function pointers
91 struct amdgpu_ring *ring; local
122 ring = &adev->vcn.inst->ring_dec;
123 ring->vm_hub = AMDGPU_MMHUB0(0);
124 sprintf(ring->name, "vcn_dec");
125 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst->irq, 0,
144 ring = &adev->vcn.inst->ring_enc[i];
145 ring->vm_hub = AMDGPU_MMHUB0(0);
146 sprintf(ring
200 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
791 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
965 struct amdgpu_ring *ring = &adev->vcn.inst->ring_dec; local
1217 struct amdgpu_ring *ring; local
1382 vcn_v1_0_dec_ring_get_rptr(struct amdgpu_ring *ring) argument
1396 vcn_v1_0_dec_ring_get_wptr(struct amdgpu_ring *ring) argument
1410 vcn_v1_0_dec_ring_set_wptr(struct amdgpu_ring *ring) argument
1428 vcn_v1_0_dec_ring_insert_start(struct amdgpu_ring *ring) argument
1447 vcn_v1_0_dec_ring_insert_end(struct amdgpu_ring *ring) argument
1466 vcn_v1_0_dec_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
1507 vcn_v1_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
1530 vcn_v1_0_dec_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1550 vcn_v1_0_dec_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
1565 vcn_v1_0_dec_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1588 vcn_v1_0_enc_ring_get_rptr(struct amdgpu_ring *ring) argument
1605 vcn_v1_0_enc_ring_get_wptr(struct amdgpu_ring *ring) argument
1622 vcn_v1_0_enc_ring_set_wptr(struct amdgpu_ring *ring) argument
1644 vcn_v1_0_enc_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
1656 vcn_v1_0_enc_ring_insert_end(struct amdgpu_ring *ring) argument
1671 vcn_v1_0_enc_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
1685 vcn_v1_0_enc_ring_emit_reg_wait(struct amdgpu_ring *ring, uint32_t reg, uint32_t val, uint32_t mask) argument
1695 vcn_v1_0_enc_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned int vmid, uint64_t pd_addr) argument
1708 vcn_v1_0_enc_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
1749 vcn_v1_0_dec_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
1828 vcn_v1_0_ring_begin_use(struct amdgpu_ring *ring) argument
1842 vcn_v1_0_set_pg_for_begin_use(struct amdgpu_ring *ring, bool set_clocks) argument
1881 vcn_v1_0_ring_end_use(struct amdgpu_ring *ring) argument
[all...]
H A Duvd_v3_1.c40 * @ring: amdgpu_ring pointer
44 static uint64_t uvd_v3_1_ring_get_rptr(struct amdgpu_ring *ring) argument
46 struct amdgpu_device *adev = ring->adev;
54 * @ring: amdgpu_ring pointer
58 static uint64_t uvd_v3_1_ring_get_wptr(struct amdgpu_ring *ring) argument
60 struct amdgpu_device *adev = ring->adev;
68 * @ring: amdgpu_ring pointer
72 static void uvd_v3_1_ring_set_wptr(struct amdgpu_ring *ring) argument
74 struct amdgpu_device *adev = ring->adev;
76 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring
89 uvd_v3_1_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
110 uvd_v3_1_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
139 uvd_v3_1_ring_test_ring(struct amdgpu_ring *ring) argument
167 uvd_v3_1_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
322 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
547 struct amdgpu_ring *ring; local
631 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
[all...]
H A Duvd_v4_2.c54 * @ring: amdgpu_ring pointer
58 static uint64_t uvd_v4_2_ring_get_rptr(struct amdgpu_ring *ring) argument
60 struct amdgpu_device *adev = ring->adev;
68 * @ring: amdgpu_ring pointer
72 static uint64_t uvd_v4_2_ring_get_wptr(struct amdgpu_ring *ring) argument
74 struct amdgpu_device *adev = ring->adev;
82 * @ring: amdgpu_ring pointer
86 static void uvd_v4_2_ring_set_wptr(struct amdgpu_ring *ring) argument
88 struct amdgpu_device *adev = ring->adev;
90 WREG32(mmUVD_RBC_RB_WPTR, lower_32_bits(ring
106 struct amdgpu_ring *ring; local
157 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
285 struct amdgpu_ring *ring = &adev->uvd.inst->ring; local
477 uvd_v4_2_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
506 uvd_v4_2_ring_test_ring(struct amdgpu_ring *ring) argument
544 uvd_v4_2_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
555 uvd_v4_2_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
[all...]
H A Dcik_sdma.c86 * and each one supports 1 ring buffer used for gfx
90 * (ring buffer, IBs, etc.), but sDMA has it's own
155 * @ring: amdgpu ring pointer
159 static uint64_t cik_sdma_ring_get_rptr(struct amdgpu_ring *ring) argument
163 rptr = *ring->rptr_cpu_addr;
171 * @ring: amdgpu ring pointer
175 static uint64_t cik_sdma_ring_get_wptr(struct amdgpu_ring *ring) argument
177 struct amdgpu_device *adev = ring
189 cik_sdma_ring_set_wptr(struct amdgpu_ring *ring) argument
197 cik_sdma_ring_insert_nop(struct amdgpu_ring *ring, uint32_t count) argument
220 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) argument
245 cik_sdma_ring_emit_hdp_flush(struct amdgpu_ring *ring) argument
276 cik_sdma_ring_emit_fence(struct amdgpu_ring *ring, u64 addr, u64 seq, unsigned flags) argument
426 struct amdgpu_ring *ring; local
598 cik_sdma_ring_test_ring(struct amdgpu_ring *ring) argument
650 cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout) argument
796 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) argument
820 cik_sdma_ring_emit_pipeline_sync(struct amdgpu_ring *ring) argument
847 cik_sdma_ring_emit_vm_flush(struct amdgpu_ring *ring, unsigned vmid, uint64_t pd_addr) argument
863 cik_sdma_ring_emit_wreg(struct amdgpu_ring *ring, uint32_t reg, uint32_t val) argument
941 struct amdgpu_ring *ring; local
[all...]
/linux-master/drivers/net/can/spi/mcp251xfd/
H A Dmcp251xfd-chip-fifo.c21 const struct mcp251xfd_rx_ring *ring)
32 ring->obj_num - 1) |
45 MCP251XFD_REG_FIFOCON(ring->fifo_nr), fifo_con);
50 const struct mcp251xfd_rx_ring *ring)
54 fltcon = MCP251XFD_REG_FLTCON_FLTEN(ring->nr) |
55 MCP251XFD_REG_FLTCON_FBP(ring->nr, ring->fifo_nr);
58 MCP251XFD_REG_FLTCON(ring->nr >> 2),
59 MCP251XFD_REG_FLTCON_FLT_MASK(ring->nr),
20 mcp251xfd_chip_rx_fifo_init_one(const struct mcp251xfd_priv *priv, const struct mcp251xfd_rx_ring *ring) argument
49 mcp251xfd_chip_rx_filter_init_one(const struct mcp251xfd_priv *priv, const struct mcp251xfd_rx_ring *ring) argument
/linux-master/drivers/mailbox/
H A Dbcm-flexrm-mailbox.c8 * extension to Broadcom FlexRM ring manager. The FlexRM ring
13 * rings where each mailbox channel represents a separate FlexRM ring.
128 /* ====== FlexRM ring descriptor defines ===== */
292 /* ====== FlexRM ring descriptor helper routines ===== */
924 struct flexrm_ring *ring; local
931 ring = &mbox->rings[i];
932 if (readl(ring->regs + RING_CONTROL) &
939 ring->num, state,
940 (unsigned long long)ring
952 struct flexrm_ring *ring; local
975 flexrm_new_request(struct flexrm_ring *ring, struct brcm_message *batch_msg, struct brcm_message *msg) argument
1074 flexrm_process_completions(struct flexrm_ring *ring) argument
1196 struct flexrm_ring *ring = chan->con_priv; local
1229 struct flexrm_ring *ring = chan->con_priv; local
1364 struct flexrm_ring *ring = chan->con_priv; local
1449 struct flexrm_ring *ring; local
1477 struct flexrm_ring *ring = &mbox->rings[desc->msi_index]; local
1491 struct flexrm_ring *ring; local
[all...]
/linux-master/net/9p/
H A Dtrans_xen.c24 #define XEN_9PFS_RING_SIZE(ring) XEN_FLEX_RING_SIZE(ring->intf->ring_order)
34 /* One per ring, more than one per 9pfs share */
42 /* protect a ring from concurrent accesses */
103 static bool p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size) argument
107 cons = ring->intf->out_cons;
108 prod = ring->intf->out_prod;
111 return XEN_9PFS_RING_SIZE(ring) -
112 xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE(ring)) >= size;
121 struct xen_9pfs_dataring *ring; local
173 struct xen_9pfs_dataring *ring; local
243 struct xen_9pfs_dataring *ring = r; local
282 struct xen_9pfs_dataring *ring = &priv->rings[i]; local
319 xen_9pfs_front_alloc_dataring(struct xenbus_device *dev, struct xen_9pfs_dataring *ring, unsigned int order) argument
[all...]
/linux-master/drivers/gpu/drm/radeon/
H A Dvce_v1_0.c56 * @ring: radeon_ring pointer
61 struct radeon_ring *ring)
63 if (ring->idx == TN_RING_TYPE_VCE1_INDEX)
73 * @ring: radeon_ring pointer
78 struct radeon_ring *ring)
80 if (ring->idx == TN_RING_TYPE_VCE1_INDEX)
90 * @ring: radeon_ring pointer
95 struct radeon_ring *ring)
97 if (ring->idx == TN_RING_TYPE_VCE1_INDEX)
98 WREG32(VCE_RB_WPTR, ring
60 vce_v1_0_get_rptr(struct radeon_device *rdev, struct radeon_ring *ring) argument
77 vce_v1_0_get_wptr(struct radeon_device *rdev, struct radeon_ring *ring) argument
94 vce_v1_0_set_wptr(struct radeon_device *rdev, struct radeon_ring *ring) argument
291 struct radeon_ring *ring; local
358 struct radeon_ring *ring; local
[all...]
/linux-master/drivers/gpu/drm/msm/adreno/
H A Da5xx_preempt.c39 /* Write the most recent wptr for the given ring into the hardware */
40 static inline void update_wptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring) argument
45 if (!ring)
48 spin_lock_irqsave(&ring->preempt_lock, flags);
49 wptr = get_wptr(ring);
50 spin_unlock_irqrestore(&ring->preempt_lock, flags);
63 struct msm_ringbuffer *ring = gpu->rb[i]; local
65 spin_lock_irqsave(&ring->preempt_lock, flags);
66 empty = (get_wptr(ring) == gpu->funcs->get_rptr(gpu, ring));
95 struct msm_ringbuffer *ring; local
220 preempt_init_ring(struct a5xx_gpu *a5xx_gpu, struct msm_ringbuffer *ring) argument
[all...]

Completed in 248 milliseconds

1234567891011>>