/linux-master/drivers/bus/mhi/host/ |
H A D | main.c | 118 struct mhi_ring *ring = &mhi_event->ring; local 121 ring->db_addr, le64_to_cpu(*ring->ctxt_wp)); 127 struct mhi_ring *ring = &mhi_cmd->ring; local 129 db = ring->iommu_base + (ring->wp - ring->base); 130 *ring 137 struct mhi_ring *ring = &mhi_chan->tre_ring; local 230 get_nr_avail_ring_elements(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring) argument 246 mhi_to_virtual(struct mhi_ring *ring, dma_addr_t addr) argument 251 mhi_add_ring_element(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring) argument 261 mhi_del_ring_element(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring) argument 271 is_valid_ring_ptr(struct mhi_ring *ring, dma_addr_t addr) argument 547 mhi_recycle_ev_ring_element(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring) argument 1109 mhi_is_ring_full(struct mhi_controller *mhi_cntrl, struct mhi_ring *ring) argument 1288 struct mhi_ring *ring = &mhi_cmd->ring; local [all...] |
/linux-master/drivers/gpu/drm/xen/ |
H A D | xen_drm_front_evtchnl.h | 17 #include <xen/interface/io/ring.h> 52 struct xen_displif_front_ring ring; member in struct:xen_drm_front_evtchnl::__anon3771::__anon3772
|
/linux-master/drivers/net/wireless/broadcom/b43legacy/ |
H A D | dma.h | 123 /* Kernel virtual base address of the ring memory. */ 129 * This is NULL for an RX ring. 132 /* (Unadjusted) DMA base bus-address of the ring memory. */ 134 /* Number of descriptor slots in the ring. */ 138 /* Currently used slot in the ring. */ 148 /* Boolean. Is this a TX ring? */ 152 /* Boolean. Is this ring stopped at ieee80211 level? */ 154 /* The QOS priority assigned to this ring. Only used for TX rings. 161 /* Last time we injected a ring overflow. */ 168 u32 b43legacy_dma_read(struct b43legacy_dmaring *ring, argument 175 b43legacy_dma_write(struct b43legacy_dmaring *ring, u16 offset, u32 value) argument 219 b43legacy_dma_rx(struct b43legacy_dmaring *ring) argument [all...] |
/linux-master/net/dccp/ccids/lib/ |
H A D | loss_interval.c | 24 return lh->counter ? lh->ring[LIH_INDEX(lh->counter - 1)] : NULL; 31 return lh->ring[LIH_INDEX(lh->counter - i - 1)]->li_length; 39 if (lh->ring[LIH_INDEX(lh->counter)] == NULL) 40 lh->ring[LIH_INDEX(lh->counter)] = kmem_cache_alloc(tfrc_lh_slab, 42 return lh->ring[LIH_INDEX(lh->counter)]; 51 if (lh->ring[LIH_INDEX(lh->counter)] != NULL) { 53 lh->ring[LIH_INDEX(lh->counter)]); 54 lh->ring[LIH_INDEX(lh->counter)] = NULL;
|
H A D | packet_history.h | 76 * @ring: Packet history for RTT sampling and loss detection 82 struct tfrc_rx_hist_entry *ring[TFRC_NDUPACK + 1]; member in struct:tfrc_rx_hist 102 return h->ring[tfrc_rx_hist_index(h, h->loss_count)]; 111 return h->ring[tfrc_rx_hist_index(h, n)]; 120 return h->ring[h->loss_start];
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | hdp_v5_0.c | 32 struct amdgpu_ring *ring) 34 if (!ring || !ring->funcs->emit_wreg) 37 amdgpu_ring_emit_wreg(ring, (adev->rmmio_remap.reg_offset + KFD_MMIO_REMAP_HDP_MEM_FLUSH_CNTL) >> 2, 0); 41 struct amdgpu_ring *ring) 43 if (!ring || !ring->funcs->emit_wreg) { 46 amdgpu_ring_emit_wreg(ring, SOC15_REG_OFFSET( 31 hdp_v5_0_flush_hdp(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 40 hdp_v5_0_invalidate_hdp(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
|
H A D | hdp_v4_0.c | 41 struct amdgpu_ring *ring) 43 if (!ring || !ring->funcs->emit_wreg) 46 amdgpu_ring_emit_wreg(ring, (adev->rmmio_remap.reg_offset + KFD_MMIO_REMAP_HDP_MEM_FLUSH_CNTL) >> 2, 0); 50 struct amdgpu_ring *ring) 56 if (!ring || !ring->funcs->emit_wreg) 59 amdgpu_ring_emit_wreg(ring, SOC15_REG_OFFSET( 40 hdp_v4_0_flush_hdp(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument 49 hdp_v4_0_invalidate_hdp(struct amdgpu_device *adev, struct amdgpu_ring *ring) argument
|
H A D | vcn_v2_5.c | 78 * Set ring and irq function pointers 123 struct amdgpu_ring *ring; local 184 ring = &adev->vcn.inst[j].ring_dec; 185 ring->use_doorbell = true; 187 ring->doorbell_index = (adev->doorbell_index.vcn.vcn_ring0_1 << 1) + 191 ring->vm_hub = AMDGPU_MMHUB1(0); 193 ring->vm_hub = AMDGPU_MMHUB0(0); 195 sprintf(ring->name, "vcn_dec_%d", j); 196 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst[j].irq, 204 ring 293 struct amdgpu_ring *ring; local 825 struct amdgpu_ring *ring; local 968 struct amdgpu_ring *ring; local 1217 struct amdgpu_ring *ring; local 1454 struct amdgpu_ring *ring; local 1533 vcn_v2_5_dec_ring_get_rptr(struct amdgpu_ring *ring) argument 1547 vcn_v2_5_dec_ring_get_wptr(struct amdgpu_ring *ring) argument 1564 vcn_v2_5_dec_ring_set_wptr(struct amdgpu_ring *ring) argument 1613 vcn_v2_5_enc_ring_get_rptr(struct amdgpu_ring *ring) argument 1630 vcn_v2_5_enc_ring_get_wptr(struct amdgpu_ring *ring) argument 1654 vcn_v2_5_enc_ring_set_wptr(struct amdgpu_ring *ring) argument [all...] |
H A D | amdgpu_sdma.c | 37 struct amdgpu_sdma_instance *amdgpu_sdma_get_instance_from_ring(struct amdgpu_ring *ring) argument 39 struct amdgpu_device *adev = ring->adev; 43 if (ring == &adev->sdma.instance[i].ring || 44 ring == &adev->sdma.instance[i].page) 50 int amdgpu_sdma_get_index_from_ring(struct amdgpu_ring *ring, uint32_t *index) argument 52 struct amdgpu_device *adev = ring->adev; 56 if (ring == &adev->sdma.instance[i].ring || 57 ring 66 amdgpu_sdma_get_csa_mc_addr(struct amdgpu_ring *ring, unsigned int vmid) argument [all...] |
H A D | amdgpu_jpeg.h | 107 void amdgpu_jpeg_ring_begin_use(struct amdgpu_ring *ring); 108 void amdgpu_jpeg_ring_end_use(struct amdgpu_ring *ring); 110 int amdgpu_jpeg_dec_ring_test_ring(struct amdgpu_ring *ring); 111 int amdgpu_jpeg_dec_ring_test_ib(struct amdgpu_ring *ring, long timeout);
|
H A D | amdgpu_mes_ctx.h | 61 uint8_t ring[PAGE_SIZE * 4]; member in struct:amdgpu_mes_ctx_meta_data::__anon37 78 uint8_t ring[PAGE_SIZE * 4]; member in struct:amdgpu_mes_ctx_meta_data::__anon38 92 uint8_t ring[PAGE_SIZE * 4]; member in struct:amdgpu_mes_ctx_meta_data::__anon39
|
/linux-master/drivers/usb/cdns3/ |
H A D | cdnsp-trace.h | 337 TP_PROTO(struct cdnsp_ring *ring, struct cdnsp_generic_trb *trb), 338 TP_ARGS(ring, trb), 349 __entry->type = ring->type; 355 __entry->trb_dma = cdnsp_trb_virt_to_dma(ring->deq_seg, 368 TP_PROTO(struct cdnsp_ring *ring, struct cdnsp_generic_trb *trb), 369 TP_ARGS(ring, trb) 373 TP_PROTO(struct cdnsp_ring *ring, struct cdnsp_generic_trb *trb), 374 TP_ARGS(ring, trb) 378 TP_PROTO(struct cdnsp_ring *ring, struct cdnsp_generic_trb *trb), 379 TP_ARGS(ring, tr [all...] |
/linux-master/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_io.c | 506 * the descriptor ring */ 939 * If everything is freed up to consumer then check if the ring is full 940 * If the ring is full then check if more needs to be freed and 944 * other filling it. If the ring is full when we get out of here and 1138 struct qlcnic_host_rds_ring *ring, 1144 buffer = &ring->rx_buf_arr[index]; 1150 dma_unmap_single(&adapter->pdev->dev, buffer->dma, ring->dma_size, 1194 struct qlcnic_host_sds_ring *sds_ring, int ring, 1205 if (unlikely(ring >= adapter->max_rds_rings)) 1208 rds_ring = &recv_ctx->rds_rings[ring]; 1137 qlcnic_process_rxbuf(struct qlcnic_adapter *adapter, struct qlcnic_host_rds_ring *ring, u16 index, u16 cksum) argument 1193 qlcnic_process_rcv(struct qlcnic_adapter *adapter, struct qlcnic_host_sds_ring *sds_ring, int ring, u64 sts_data0) argument 1261 qlcnic_process_lro(struct qlcnic_adapter *adapter, int ring, u64 sts_data0, u64 sts_data1) argument 1365 u8 ring; local 1491 qlcnic_process_rcv_diag(struct qlcnic_adapter *adapter, int ring, u64 sts_data0) argument 1541 int ring, opcode, desc_cnt; local 1576 int ring; local 1618 int ring; local 1642 int ring; local 1669 int ring; local 1715 qlcnic_83xx_process_rcv(struct qlcnic_adapter *adapter, struct qlcnic_host_sds_ring *sds_ring, u8 ring, u64 sts_data[]) argument 1782 qlcnic_83xx_process_lro(struct qlcnic_adapter *adapter, u8 ring, u64 sts_data[]) argument 1886 u8 ring; local 2045 int ring; local 2072 int ring; local 2102 int ring; local 2145 int ring; local 2168 qlcnic_83xx_process_rcv_diag(struct qlcnic_adapter *adapter, int ring, u64 sts_data[]) argument 2209 int ring, opcode; local [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_test.c | 41 int i, r, ring; local 45 ring = radeon_copy_dma_ring_index(rdev); 48 ring = radeon_copy_blit_ring_index(rdev); 58 * (Total GTT - IB pool - writeback page - ring buffers) / test size 120 if (ring == R600_RING_TYPE_DMA_INDEX) 171 if (ring == R600_RING_TYPE_DMA_INDEX) 263 struct radeon_ring *ring, 266 uint32_t handle = ring->idx ^ 0xdeafbeef; 269 if (ring->idx == R600_RING_TYPE_UVD_INDEX) { 270 r = radeon_uvd_get_create_msg(rdev, ring 262 radeon_test_create_and_emit_fence(struct radeon_device *rdev, struct radeon_ring *ring, struct radeon_fence **fence) argument [all...] |
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_ethtool.c | 97 struct wx_ring *ring; local 111 ring = wx->tx_ring[j]; 112 if (!ring) { 119 start = u64_stats_fetch_begin(&ring->syncp); 120 data[i] = ring->stats.packets; 121 data[i + 1] = ring->stats.bytes; 122 } while (u64_stats_fetch_retry(&ring->syncp, start)); 126 ring = wx->rx_ring[j]; 127 if (!ring) { 134 start = u64_stats_fetch_begin(&ring 234 wx_get_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/drivers/net/ethernet/synopsys/ |
H A D | dwc-xlgmac-net.c | 28 static inline unsigned int xlgmac_tx_avail_desc(struct xlgmac_ring *ring) argument 30 return (ring->dma_desc_count - (ring->cur - ring->dirty)); 33 static inline unsigned int xlgmac_rx_dirty_desc(struct xlgmac_ring *ring) argument 35 return (ring->cur - ring->dirty); 40 struct xlgmac_ring *ring, 45 if (count > xlgmac_tx_avail_desc(ring)) { 49 ring 38 xlgmac_maybe_stop_tx_queue( struct xlgmac_channel *channel, struct xlgmac_ring *ring, unsigned int count) argument 114 xlgmac_prep_tx_pkt(struct xlgmac_pdata *pdata, struct xlgmac_ring *ring, struct sk_buff *skb, struct xlgmac_pkt_info *pkt_info) argument 523 struct xlgmac_ring *ring; local 544 struct xlgmac_ring *ring; local 707 struct xlgmac_ring *ring; local 953 struct xlgmac_ring *ring = channel->rx_ring; local 1036 struct xlgmac_ring *ring = channel->tx_ring; local 1109 struct xlgmac_ring *ring = channel->rx_ring; local [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_engine_user.c | 160 static int legacy_ring_idx(const struct legacy_ring *ring) argument 172 if (GEM_DEBUG_WARN_ON(ring->class >= ARRAY_SIZE(map))) 175 if (GEM_DEBUG_WARN_ON(ring->instance >= map[ring->class].max)) 178 return map[ring->class].base + ring->instance; 181 static void add_legacy_ring(struct legacy_ring *ring, argument 184 if (engine->gt != ring->gt || engine->class != ring->class) { 185 ring 207 struct legacy_ring ring = {}; local [all...] |
/linux-master/tools/testing/selftests/drivers/net/netdevsim/ |
H A D | Makefile | 9 ethtool-ring.sh \
|
H A D | ethtool-ring.sh | 21 if ! ethtool -h | grep -q set-ring >/dev/null; then 22 echo "SKIP: No --set-ring support in ethtool" 52 RING_MAX_LIST=$(ls $NSIM_DEV_DFS/ethtool/ring/) 55 echo $MAX_VALUE > $NSIM_DEV_DFS/ethtool/ring/$ring_max_entry
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3_common/ |
H A D | hclge_comm_cmd.c | 8 struct hclge_comm_cmq_ring *ring) 10 dma_addr_t dma = ring->desc_dma_addr; 13 if (ring->ring_type == HCLGE_COMM_TYPE_CSQ) { 20 reg_val |= ring->desc_num >> HCLGE_COMM_NIC_CMQ_DESC_NUM_S; 29 reg_val = ring->desc_num >> HCLGE_COMM_NIC_CMQ_DESC_NUM_S; 102 void hclge_comm_free_cmd_desc(struct hclge_comm_cmq_ring *ring) argument 104 int size = ring->desc_num * sizeof(struct hclge_desc); 106 if (!ring->desc) 109 dma_free_coherent(&ring->pdev->dev, size, 110 ring 7 hclge_comm_cmd_config_regs(struct hclge_comm_hw *hw, struct hclge_comm_cmq_ring *ring) argument 114 hclge_comm_alloc_cmd_desc(struct hclge_comm_cmq_ring *ring) argument 209 struct hclge_comm_cmq_ring *ring = local 283 hclge_comm_ring_space(struct hclge_comm_cmq_ring *ring) argument 308 hclge_comm_is_valid_csq_clean_head(struct hclge_comm_cmq_ring *ring, int head) argument [all...] |
/linux-master/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_rq.c | 21 unsigned int i, j, count = rq->ring.desc_count; 34 buf->desc = (u8 *)rq->ring.descs + 35 rq->ring.desc_size * buf->index; 60 vnic_dev_free_desc_ring(vdev, &rq->ring); 88 err = vnic_dev_alloc_desc_ring(vdev, &rq->ring, desc_count, desc_size); 107 unsigned int count = rq->ring.desc_count; 109 paddr = (u64)rq->ring.base_addr | VNIC_PADDR_TARGET; 176 unsigned int count = rq->ring.desc_count; 181 for (i = 0; i < rq->ring.desc_count; i++) { 185 rq->ring [all...] |
/linux-master/drivers/scsi/fnic/ |
H A D | vnic_wq.c | 31 return vnic_dev_alloc_desc_ring(vdev, &wq->ring, desc_count, desc_size); 38 unsigned int i, j, count = wq->ring.desc_count; 53 buf->desc = (u8 *)wq->ring.descs + 54 wq->ring.desc_size * buf->index; 79 vnic_dev_free_desc_ring(vdev, &wq->ring); 106 err = vnic_dev_alloc_desc_ring(vdev, &wq->ring, desc_count, desc_size); 147 unsigned int count = wq->ring.desc_count; 149 paddr = (u64)wq->ring.base_addr | VNIC_PADDR_TARGET; 171 paddr = (u64)wq->ring.base_addr | VNIC_PADDR_TARGET; 173 iowrite32(wq->ring [all...] |
/linux-master/net/xdp/ |
H A D | xsk_queue.c | 2 /* XDP user-space ring structure 49 q->ring = vmalloc_user(size); 50 if (!q->ring) { 64 vfree(q->ring);
|
/linux-master/drivers/bus/mhi/ep/ |
H A D | internal.h | 121 struct mhi_ep_ring *ring; member in struct:mhi_ep_ring_item 145 struct mhi_ep_ring ring; member in struct:mhi_ep_cmd 149 struct mhi_ep_ring ring; member in struct:mhi_ep_event 160 struct mhi_ep_ring ring; member in struct:mhi_ep_chan 174 void mhi_ep_ring_init(struct mhi_ep_ring *ring, enum mhi_ep_ring_type type, u32 id); 175 void mhi_ep_ring_reset(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_ring *ring); 176 int mhi_ep_ring_start(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_ring *ring, 178 size_t mhi_ep_ring_addr2offset(struct mhi_ep_ring *ring, u64 ptr); 179 int mhi_ep_ring_add_element(struct mhi_ep_ring *ring, struct mhi_ring_element *element); 180 void mhi_ep_ring_inc_index(struct mhi_ep_ring *ring); [all...] |
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_gen4_hw_data.h | 39 /* Bank and ring configuration */ 76 #define READ_CSR_RING_HEAD(csr_base_addr, bank, ring) \ 79 ADF_RING_CSR_RING_HEAD + ((ring) << 2)) 80 #define READ_CSR_RING_TAIL(csr_base_addr, bank, ring) \ 83 ADF_RING_CSR_RING_TAIL + ((ring) << 2)) 87 #define WRITE_CSR_RING_CONFIG(csr_base_addr, bank, ring, value) \ 90 ADF_RING_CSR_RING_CONFIG + ((ring) << 2), value) 91 #define WRITE_CSR_RING_BASE(csr_base_addr, bank, ring, value) \ 95 u32 _ring = ring; \ 108 #define WRITE_CSR_RING_HEAD(csr_base_addr, bank, ring, valu [all...] |