/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/enic/ |
H A D | vnic_rq.c | 34 unsigned int i, j, count = rq->ring.desc_count; 51 buf->desc = (u8 *)rq->ring.descs + 52 rq->ring.desc_size * buf->index; 77 vnic_dev_free_desc_ring(vdev, &rq->ring); 103 err = vnic_dev_alloc_desc_ring(vdev, &rq->ring, desc_count, desc_size); 122 unsigned int count = rq->ring.desc_count; 124 paddr = (u64)rq->ring.base_addr | VNIC_PADDR_TARGET; 146 /* Use current fetch_index as the ring starting point */ 193 unsigned int count = rq->ring.desc_count; 202 rq->ring [all...] |
H A D | vnic_wq.h | 79 struct vnic_dev_ring ring; member in struct:vnic_wq 89 return wq->ring.desc_avail; 95 return wq->ring.desc_count - wq->ring.desc_avail - 1; 126 wq->ring.desc_avail--; 142 wq->ring.desc_avail++;
|
H A D | vnic_dev.c | 193 unsigned int vnic_dev_desc_ring_size(struct vnic_dev_ring *ring, argument 205 ring->base_align = 512; 210 ring->desc_count = ALIGN(desc_count, count_align); 212 ring->desc_size = ALIGN(desc_size, desc_align); 214 ring->size = ring->desc_count * ring->desc_size; 215 ring->size_unaligned = ring->size + ring 220 vnic_dev_clear_desc_ring(struct vnic_dev_ring *ring) argument 225 vnic_dev_alloc_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring, unsigned int desc_count, unsigned int desc_size) argument 252 vnic_dev_free_desc_ring(struct vnic_dev *vdev, struct vnic_dev_ring *ring) argument [all...] |
H A D | vnic_wq.c | 34 unsigned int i, j, count = wq->ring.desc_count; 51 buf->desc = (u8 *)wq->ring.descs + 52 wq->ring.desc_size * buf->index; 77 vnic_dev_free_desc_ring(vdev, &wq->ring); 103 err = vnic_dev_alloc_desc_ring(vdev, &wq->ring, desc_count, desc_size); 122 unsigned int count = wq->ring.desc_count; 124 paddr = (u64)wq->ring.base_addr | VNIC_PADDR_TARGET; 188 wq->ring.desc_avail++; 197 vnic_dev_clear_desc_ring(&wq->ring);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/net/dccp/ccids/lib/ |
H A D | packet_history.c | 191 struct tfrc_rx_hist_entry *tmp = h->ring[idx_a]; 193 h->ring[idx_a] = h->ring[idx_b]; 194 h->ring[idx_b] = tmp; 392 h->ring[i] = kmem_cache_alloc(tfrc_rx_hist_slab, GFP_ATOMIC); 393 if (h->ring[i] == NULL) 402 kmem_cache_free(tfrc_rx_hist_slab, h->ring[i]); 403 h->ring[i] = NULL; 413 if (h->ring[i] != NULL) { 414 kmem_cache_free(tfrc_rx_hist_slab, h->ring[ [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/netxen/ |
H A D | netxen_nic_ctx.c | 553 int ring; local 566 for (ring = 0; ring < adapter->max_rds_rings; ring++) { 567 rds_ring = &recv_ctx->rds_rings[ring]; 569 hwctx->rcv_rings[ring].addr = 571 hwctx->rcv_rings[ring].size = 575 for (ring = 0; ring < adapter->max_sds_rings; ring 604 int ring; local 725 int ring; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/fnic/ |
H A D | vnic_wq.h | 89 struct vnic_dev_ring ring; member in struct:vnic_wq 99 return wq->ring.desc_avail; 105 return wq->ring.desc_count - wq->ring.desc_avail - 1; 136 wq->ring.desc_avail--; 152 wq->ring.desc_avail++;
|
H A D | vnic_rq.c | 31 unsigned int i, j, count = rq->ring.desc_count; 48 buf->desc = (u8 *)rq->ring.descs + 49 rq->ring.desc_size * buf->index; 75 vnic_dev_free_desc_ring(vdev, &rq->ring); 101 err = vnic_dev_alloc_desc_ring(vdev, &rq->ring, desc_count, desc_size); 121 paddr = (u64)rq->ring.base_addr | VNIC_PADDR_TARGET; 123 iowrite32(rq->ring.desc_count, &rq->ctrl->ring_size); 130 /* Use current fetch_index as the ring starting point */ 183 rq->ring.desc_avail++; 186 /* Use current fetch_index as the ring startin [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/ |
H A D | pasemi_mac.c | 316 struct pasemi_mac_csring *ring; local 321 ring = pasemi_dma_alloc_chan(TXCHAN, sizeof(struct pasemi_mac_csring), 324 if (!ring) { 329 chno = ring->chan.chno; 331 ring->size = CS_RING_SIZE; 332 ring->next_to_fill = 0; 335 if (pasemi_dma_alloc_ring(&ring->chan, CS_RING_SIZE)) 339 PAS_DMA_TXCHAN_BASEL_BRBL(ring->chan.ring_dma)); 340 val = PAS_DMA_TXCHAN_BASEU_BRBH(ring->chan.ring_dma >> 32); 345 ring 414 struct pasemi_mac_rxring *ring; local 499 struct pasemi_mac_txring *ring; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/staging/iio/ |
H A D | Makefile | 7 industrialio-$(CONFIG_IIO_RING_BUFFER) += industrialio-ring.o
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/staging/iio/imu/ |
H A D | adis16300.h | 119 /* At the moment triggers are only used for ring buffer 144 int adis16300_initialize_ring(struct iio_ring_buffer *ring); 145 void adis16300_uninitialize_ring(struct iio_ring_buffer *ring); 174 static inline int adis16300_initialize_ring(struct iio_ring_buffer *ring) argument 179 static inline void adis16300_uninitialize_ring(struct iio_ring_buffer *ring) argument
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/usb/host/ |
H A D | xhci-ring.c | 28 * Consumer Cycle State (CCS), depending on ring function. 32 * 1. A ring is empty if enqueue == dequeue. This means there will always be at 33 * least one free TRB in the ring. This is useful if you want to turn that 34 * into a link TRB and expand the ring. 37 * link TRB had its toggle bit set, you may need to update the ring cycle 40 * 3. A ring is full if enqueue++ (for the definition of increment above) 45 * in a link TRB, it must toggle the ring cycle state. 47 * in a link TRB, it must toggle the ring cycle state. 50 * 1. Check if ring is full before you enqueue. 51 * 2. Write the ring cycl 96 last_trb_on_last_seg(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_segment *seg, union xhci_trb *trb) argument 110 last_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_segment *seg, union xhci_trb *trb) argument 119 enqueue_is_link_trb(struct xhci_ring *ring) argument 129 next_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, struct xhci_segment **seg, union xhci_trb **trb) argument 146 inc_deq(struct xhci_hcd *xhci, struct xhci_ring *ring, bool consumer) argument 203 inc_enq(struct xhci_hcd *xhci, struct xhci_ring *ring, bool consumer, bool more_trbs_coming) argument 265 room_on_ring(struct xhci_hcd *xhci, struct xhci_ring *ring, unsigned int num_trbs) argument 778 struct xhci_ring *ring; local 2136 queue_trb(struct xhci_hcd *xhci, struct xhci_ring *ring, bool consumer, bool more_trbs_coming, u32 field1, u32 field2, u32 field3, u32 field4) argument 2181 struct xhci_ring *ring = ep_ring; local [all...] |
H A D | u132-hcd.c | 150 struct u132_ring *ring; member in struct:u132_endp 185 struct u132_ring ring[MAX_U132_RINGS]; member in struct:u132 300 static inline void u132_ring_put_kref(struct u132 *u132, struct u132_ring *ring) argument 305 static void u132_ring_requeue_work(struct u132 *u132, struct u132_ring *ring, argument 309 if (queue_delayed_work(workqueue, &ring->scheduler, delta)) 311 } else if (queue_delayed_work(workqueue, &ring->scheduler, 0)) 317 static void u132_ring_queue_work(struct u132 *u132, struct u132_ring *ring, argument 321 u132_ring_requeue_work(u132, ring, delta); 325 static void u132_ring_cancel_work(struct u132 *u132, struct u132_ring *ring) argument 327 if (cancel_delayed_work(&ring 341 struct u132_ring *ring = endp->ring; local 512 struct u132_ring *ring; local 576 edset_input(struct u132 *u132, struct u132_ring *ring, struct u132_endp *endp, struct urb *urb, u8 address, u8 toggle_bits, void (*callback) (void *endp, struct urb *urb, u8 *buf, int len, int toggle_bits, int error_count, int condition_code, int repeat_number, int halted, int skipped, int actual, int non_null)) argument 586 edset_setup(struct u132 *u132, struct u132_ring *ring, struct u132_endp *endp, struct urb *urb, u8 address, u8 toggle_bits, void (*callback) (void *endp, struct urb *urb, u8 *buf, int len, int toggle_bits, int error_count, int condition_code, int repeat_number, int halted, int skipped, int actual, int non_null)) argument 596 edset_single(struct u132 *u132, struct u132_ring *ring, struct u132_endp *endp, struct urb *urb, u8 address, u8 toggle_bits, void (*callback) (void *endp, struct urb *urb, u8 *buf, int len, int toggle_bits, int error_count, int condition_code, int repeat_number, int halted, int skipped, int actual, int non_null)) argument 606 edset_output(struct u132 *u132, struct u132_ring *ring, struct u132_endp *endp, struct urb *urb, u8 address, u8 toggle_bits, void (*callback) (void *endp, struct urb *urb, u8 *buf, int len, int toggle_bits, int error_count, int condition_code, int repeat_number, int halted, int skipped, int actual, int non_null)) argument 648 struct u132_ring *ring = endp->ring; local 747 struct u132_ring *ring = endp->ring; local 799 struct u132_ring *ring = endp->ring; local 938 struct u132_ring *ring = endp->ring; local 1049 struct u132_ring *ring = endp->ring; local 1060 struct u132_ring *ring = endp->ring; local 1146 struct u132_ring *ring = endp->ring; local 1227 struct u132_ring *ring = endp->ring; local 1279 struct u132_ring *ring = endp->ring; local 1302 struct u132_ring *ring = local 1364 struct u132_ring *ring; local 1865 struct u132_ring *ring; local 1964 struct u132_ring *ring; local 2060 struct u132_ring *ring; local 3009 struct u132_ring *ring = &u132->ring[rings]; local 3042 struct u132_ring *ring = &u132->ring[rings]; local [all...] |
H A D | xhci-dbg.c | 83 xhci_dbg(xhci, " Maximum allowed segments in event ring: %u\n", 128 xhci_dbg(xhci, " Event ring is %sempty\n", 258 xhci_dbg(xhci, "Next ring segment DMA address = 0x%llx\n", address); 305 void xhci_dbg_ring_ptrs(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 308 ring->dequeue, 309 (unsigned long long)xhci_trb_virt_to_dma(ring->deq_seg, 310 ring->dequeue)); 312 ring->deq_updates); 314 ring->enqueue, 315 (unsigned long long)xhci_trb_virt_to_dma(ring 330 xhci_debug_ring(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 349 struct xhci_ring *ring; local [all...] |
H A D | xhci-mem.c | 31 * Allocates a generic ring segment from the ring pool, sets the dma address, 108 void xhci_ring_free(struct xhci_hcd *xhci, struct xhci_ring *ring) argument 113 if (!ring || !ring->first_seg) 115 first_seg = ring->first_seg; 117 xhci_dbg(xhci, "Freeing ring at %p\n", ring); 124 ring->first_seg = NULL; 125 kfree(ring); 128 xhci_initialize_ring_info(struct xhci_ring *ring) argument 155 struct xhci_ring *ring; local 229 xhci_reinit_cached_ring(struct xhci_hcd *xhci, struct xhci_ring *ring) argument [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/vxge/ |
H A D | vxge-main.c | 126 struct vxge_ring *ring; local 130 ring = &vdev->vpaths[i].ring; 131 vxge_hw_vpath_poll_rx(ring->handle); 189 vxge_rx_alloc(void *dtrh, struct vxge_ring *ring, const int skb_size) argument 195 dev = ring->ndev; 197 ring->ndev->name, __func__, __LINE__); 207 ring->stats.skb_alloc_fail++; 212 "%s: %s:%d Skb : 0x%p", ring->ndev->name, 221 "%s: %s:%d Exiting...", ring 229 vxge_rx_map(void *dtrh, struct vxge_ring *ring) argument 266 struct vxge_ring *ring = (struct vxge_ring *)userdata; local 288 vxge_rx_complete(struct vxge_ring *ring, struct sk_buff *skb, u16 vlan, int pkt_length, struct vxge_hw_ring_rxd_info *ext_info) argument 327 vxge_re_pre_post(void *dtr, struct vxge_ring *ring, struct vxge_rx_priv *rx_priv) argument 361 struct vxge_ring *ring = (struct vxge_ring *)userdata; local 938 struct vxge_ring *ring = (struct vxge_ring *)userdata; local 1575 struct vxge_ring *ring = local 1599 struct vxge_ring *ring; local 2111 struct vxge_ring *ring = (struct vxge_ring *)dev_id; local 2476 struct vxge_ring *ring; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/gpu/drm/i830/ |
H A D | i830_dma.c | 73 * yet, the hardware updates in use for us once its on the ring buffer. 225 if (dev_priv->ring.virtual_start) 226 drm_core_ioremapfree(&dev_priv->ring.map, dev); 251 drm_i830_ring_buffer_t *ring = &(dev_priv->ring); local 257 while (ring->space < n) { 258 ring->head = I830_READ(LP_RING + RING_HEAD) & HEAD_ADDR; 259 ring->space = ring->head - (ring 285 drm_i830_ring_buffer_t *ring = &(dev_priv->ring); local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/wireless/rtl818x/ |
H A D | rtl8180_dev.c | 176 struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; local 178 while (skb_queue_len(&ring->queue)) { 179 struct rtl8180_tx_desc *entry = &ring->desc[ring->idx]; 187 ring->idx = (ring->idx + 1) % ring->entries; 188 skb = __skb_dequeue(&ring->queue); 203 if (ring->entries - skb_queue_len(&ring 248 struct rtl8180_tx_ring *ring; local 499 struct rtl8180_tx_desc *ring; local 527 struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/igb/ |
H A D | igb.h | 192 dma_addr_t dma; /* phys address of the ring */ 193 void *desc; /* descriptor ring memory */ 194 unsigned int size; /* length of desc. ring in bytes */ 195 u16 count; /* number of desc. in the ring */ 238 static inline int igb_desc_unused(struct igb_ring *ring) argument 240 if (ring->next_to_clean > ring->next_to_use) 241 return ring->next_to_clean - ring->next_to_use - 1; 243 return ring [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/usb/musb/ |
H A D | musb_host.h | 54 struct list_head ring; /* of musb_qh */ member in struct:musb_qh 76 /* map from control or bulk queue head to the first qh on that ring */ 81 return list_entry(q->next, struct musb_qh, ring);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/igbvf/ |
H A D | ethtool.c | 261 struct ethtool_ringparam *ring) 267 ring->rx_max_pending = IGBVF_MAX_RXD; 268 ring->tx_max_pending = IGBVF_MAX_TXD; 269 ring->rx_mini_max_pending = 0; 270 ring->rx_jumbo_max_pending = 0; 271 ring->rx_pending = rx_ring->count; 272 ring->tx_pending = tx_ring->count; 273 ring->rx_mini_pending = 0; 274 ring->rx_jumbo_pending = 0; 278 struct ethtool_ringparam *ring) 260 igbvf_get_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring) argument 277 igbvf_set_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring) argument [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/staging/iio/accel/ |
H A D | lis3l02dq_ring.c | 143 * lis3l02dq_read_accel_from_ring() individual acceleration read from ring 172 data = kmalloc(dev_info->ring->access.get_bpd(dev_info->ring), 176 ret = dev_info->ring->access.read_last(dev_info->ring, 189 len = sprintf(buf, "ring %d\n", data[len]); 359 * switched before the ring is reenabled. 475 iio_sw_rb_free(indio_dev->ring); 493 indio_dev->ring = iio_sw_rb_allocate(indio_dev); 494 if (!indio_dev->ring) [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/gpu/drm/i810/ |
H A D | i810_dma.c | 71 * yet, the hardware updates in use for us once its on the ring buffer. 221 if (dev_priv->ring.virtual_start) 222 drm_core_ioremapfree(&dev_priv->ring.map, dev); 247 drm_i810_ring_buffer_t *ring = &(dev_priv->ring); local 253 while (ring->space < n) { 254 ring->head = I810_READ(LP_RING + RING_HEAD) & HEAD_ADDR; 255 ring->space = ring->head - (ring 280 drm_i810_ring_buffer_t *ring = &(dev_priv->ring); local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/mlx4/ |
H A D | en_cq.c | 48 int entries, int ring, enum cq_type mode) 56 cq->vector = ring % mdev->dev->caps.num_comp_vectors; 62 cq->ring = ring; 93 cq->size = priv->rx_ring[cq->ring].actual_size; 46 mlx4_en_create_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, int entries, int ring, enum cq_type mode) argument
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/qlcnic/ |
H A D | qlcnic_init.c | 108 int i, ring; local 111 for (ring = 0; ring < adapter->max_rds_rings; ring++) { 112 rds_ring = &recv_ctx->rds_rings[ring]; 133 int i, ring; local 136 for (ring = 0; ring < adapter->max_rds_rings; ring++) { 137 rds_ring = &recv_ctx->rds_rings[ring]; 191 int ring; local 223 int ring, i, size; local 1369 qlcnic_process_rcv(struct qlcnic_adapter *adapter, struct qlcnic_host_sds_ring *sds_ring, int ring, u64 sts_data0) argument 1422 qlcnic_process_lro(struct qlcnic_adapter *adapter, struct qlcnic_host_sds_ring *sds_ring, int ring, u64 sts_data0, u64 sts_data1) argument 1501 int opcode, ring, desc_cnt; local 1666 qlcnic_process_rcv_diag(struct qlcnic_adapter *adapter, struct qlcnic_host_sds_ring *sds_ring, int ring, u64 sts_data0) argument 1718 int opcode, ring, desc_cnt; local [all...] |