Lines Matching refs:ring

11 					struct ath12k_dbring *ring,
23 srng = &ab->hal.srng_list[ring->refill_srng.ring_id];
30 ptr_aligned = PTR_ALIGN(ptr_unaligned, ring->buf_align);
31 paddr = dma_map_single(ab->dev, ptr_aligned, ring->buf_sz,
38 spin_lock_bh(&ring->idr_lock);
39 buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp);
40 spin_unlock_bh(&ring->idr_lock);
64 spin_lock_bh(&ring->idr_lock);
65 idr_remove(&ring->bufs_idr, buf_id);
66 spin_unlock_bh(&ring->idr_lock);
68 dma_unmap_single(ab->dev, paddr, ring->buf_sz,
76 struct ath12k_dbring *ring,
86 srng = &ab->hal.srng_list[ring->refill_srng.ring_id];
91 req_entries = min(num_free, ring->bufs_max);
93 align = ring->buf_align;
94 size = sizeof(*buff) + ring->buf_sz + align - 1;
101 ret = ath12k_dbring_bufs_replenish(ar, ring, buff, gfp);
103 ath12k_warn(ab, "failed to replenish db ring num_remain %d req_ent %d\n",
117 struct ath12k_dbring *ring,
126 arg.pdev_id = DP_SW2HW_MACID(ring->pdev_id);
128 arg.base_paddr_lo = lower_32_bits(ring->refill_srng.paddr);
129 arg.base_paddr_hi = upper_32_bits(ring->refill_srng.paddr);
130 arg.head_idx_paddr_lo = lower_32_bits(ring->hp_addr);
131 arg.head_idx_paddr_hi = upper_32_bits(ring->hp_addr);
132 arg.tail_idx_paddr_lo = lower_32_bits(ring->tp_addr);
133 arg.tail_idx_paddr_hi = upper_32_bits(ring->tp_addr);
134 arg.num_elems = ring->bufs_max;
135 arg.buf_size = ring->buf_sz;
136 arg.num_resp_per_event = ring->num_resp_per_event;
137 arg.event_timeout_ms = ring->event_timeout_ms;
141 ath12k_warn(ar->ab, "failed to setup db ring cfg\n");
148 int ath12k_dbring_set_cfg(struct ath12k *ar, struct ath12k_dbring *ring,
153 if (WARN_ON(!ring))
156 ring->num_resp_per_event = num_resp_per_event;
157 ring->event_timeout_ms = event_timeout_ms;
158 ring->handler = handler;
164 struct ath12k_dbring *ring,
171 srng = &ab->hal.srng_list[ring->refill_srng.ring_id];
172 ring->bufs_max = ring->refill_srng.size /
175 ring->buf_sz = db_cap->min_buf_sz;
176 ring->buf_align = db_cap->min_buf_align;
177 ring->pdev_id = db_cap->pdev_id;
178 ring->hp_addr = ath12k_hal_srng_get_hp_addr(ab, srng);
179 ring->tp_addr = ath12k_hal_srng_get_tp_addr(ab, srng);
181 ret = ath12k_dbring_fill_bufs(ar, ring, GFP_KERNEL);
186 int ath12k_dbring_srng_setup(struct ath12k *ar, struct ath12k_dbring *ring,
191 ret = ath12k_dp_srng_setup(ar->ab, &ring->refill_srng, HAL_RXDMA_DIR_BUF,
201 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng);
233 struct ath12k_dbring *ring = NULL;
275 ring = NULL;
281 if (!ring) {
286 srng = &ab->hal.srng_list[ring->refill_srng.ring_id];
288 size = sizeof(*buff) + ring->buf_sz + ring->buf_align - 1;
304 spin_lock_bh(&ring->idr_lock);
305 buff = idr_find(&ring->bufs_idr, buf_id);
307 spin_unlock_bh(&ring->idr_lock);
310 idr_remove(&ring->bufs_idr, buf_id);
311 spin_unlock_bh(&ring->idr_lock);
313 dma_unmap_single(ab->dev, buff->paddr, ring->buf_sz,
316 if (ring->handler) {
319 ring->buf_align);
320 handler_data.data_sz = ring->buf_sz;
322 ring->handler(ar, &handler_data);
326 ath12k_dbring_bufs_replenish(ar, ring, buff, GFP_ATOMIC);
337 void ath12k_dbring_srng_cleanup(struct ath12k *ar, struct ath12k_dbring *ring)
339 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng);
342 void ath12k_dbring_buf_cleanup(struct ath12k *ar, struct ath12k_dbring *ring)
347 spin_lock_bh(&ring->idr_lock);
348 idr_for_each_entry(&ring->bufs_idr, buff, buf_id) {
349 idr_remove(&ring->bufs_idr, buf_id);
351 ring->buf_sz, DMA_FROM_DEVICE);
355 idr_destroy(&ring->bufs_idr);
356 spin_unlock_bh(&ring->idr_lock);