Searched refs:PKT_BUF_SZ (Results 1 - 25 of 40) sorted by relevance

12

/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/
H A Dsunqe.h321 #define PKT_BUF_SZ 1664 macro
325 u8 tx_buf[TX_RING_SIZE][PKT_BUF_SZ];
327 u8 rx_buf[RX_RING_SIZE][PKT_BUF_SZ];
H A Dlance.c144 statically allocates full-sized (slightly oversized -- PKT_BUF_SZ) buffers to
201 #define PKT_BUF_SZ 1544 macro
247 char (*tx_bounce_buffs)[PKT_BUF_SZ];
543 lp->rx_buffs = (unsigned long)kmalloc(PKT_BUF_SZ*RX_RING_SIZE,
548 lp->tx_bounce_buffs = kmalloc(PKT_BUF_SZ*TX_RING_SIZE,
867 skb = alloc_skb(PKT_BUF_SZ, GFP_DMA | gfp);
873 rx_buff = kmalloc(PKT_BUF_SZ, GFP_DMA | gfp);
878 lp->rx_ring[i].buf_length = -PKT_BUF_SZ;
1201 lp->rx_ring[entry].buf_length = -PKT_BUF_SZ;
H A Dlasi_82596.c190 #define PKT_BUF_SZ 1536 macro
555 struct sk_buff *skb = dev_alloc_skb(PKT_BUF_SZ + 4);
560 dma_addr = dma_map_single(lp->dev, skb->data,PKT_BUF_SZ,
569 rbd->size = PKT_BUF_SZ;
609 PKT_BUF_SZ, DMA_FROM_DEVICE);
775 dma_unmap_single(lp->dev,(dma_addr_t)WSWAPchar(rbd->b_data), PKT_BUF_SZ, DMA_FROM_DEVICE);
777 newskb = dev_alloc_skb(PKT_BUF_SZ + 4);
789 dma_addr = dma_map_single(lp->dev, newskb->data, PKT_BUF_SZ, DMA_FROM_DEVICE);
804 dma_sync_single_for_cpu(lp->dev, (dma_addr_t)WSWAPchar(rbd->b_data), PKT_BUF_SZ, DMA_FROM_DEVICE);
807 dma_sync_single_for_device(lp->dev, (dma_addr_t)WSWAPchar(rbd->b_data), PKT_BUF_SZ, DMA_FROM_DEVIC
[all...]
H A Dsgiseeq.c45 #define PKT_BUF_SZ 1584 macro
152 #define RCNTINFO_INIT (RCNTCFG_INIT | (PKT_BUF_SZ & HPCDMA_BCNT))
170 buffer = (unsigned long) kmalloc(PKT_BUF_SZ, GFP_KERNEL);
184 buffer = (unsigned long) kmalloc(PKT_BUF_SZ, GFP_KERNEL);
310 len = PKT_BUF_SZ - (rd->rdma.cntinfo & HPCDMA_BCNT) - 3;
H A Datarilance.c241 #define PKT_BUF_SZ 1544 macro
708 if (((o) < RIEBL_RSVD_START) ? (o)+PKT_BUF_SZ > RIEBL_RSVD_START \
721 offset += PKT_BUF_SZ;
729 MEM->rx_head[i].buf_length = -PKT_BUF_SZ;
731 offset += PKT_BUF_SZ;
H A Dsun3lance.c103 #define PKT_BUF_SZ 1544 macro
144 char rx_data[RX_RING_SIZE][PKT_BUF_SZ];
145 char tx_data[TX_RING_SIZE][PKT_BUF_SZ];
477 MEM->rx_head[i].buf_length = -PKT_BUF_SZ | 0xf000;
837 // head->buf_length = -PKT_BUF_SZ | 0xf000;
H A Dgt64240eth.h86 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer. */ macro
H A Dgt64240eth.c926 PKT_BUF_SZ * RX_RING_SIZE, &gp->rx_buff_dma,
952 dma_free_noncoherent(NULL, PKT_BUF_SZ * RX_RING_SIZE,
1046 gp->rx_ring[i].buff_ptr = gp->rx_buff_dma + i * PKT_BUF_SZ;
1047 gp->rx_ring[i].buff_sz = PKT_BUF_SZ;
1545 &gp->rx_buff[nextOut * PKT_BUF_SZ], pkt_len);
H A Dpcnet32.c177 #define PKT_BUF_SZ 1544 macro
591 new_skb_list[new] = dev_alloc_skb(PKT_BUF_SZ);
604 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
606 new_rx_ring[new].buf_length = le16_to_cpu(2 - PKT_BUF_SZ);
613 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
638 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
665 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
1183 if (unlikely(pkt_len > PKT_BUF_SZ - 2)) {
1200 if ((newskb = dev_alloc_skb(PKT_BUF_SZ))) {
1205 PKT_BUF_SZ
[all...]
H A Deepro100.c88 #define PKT_BUF_SZ 1536 macro
1245 skb = dev_alloc_skb(PKT_BUF_SZ + sizeof(struct RxFD));
1256 PKT_BUF_SZ + sizeof(struct RxFD), PCI_DMA_BIDIRECTIONAL);
1269 rxf->count = cpu_to_le32(PKT_BUF_SZ << 16);
1622 skb = dev_alloc_skb(PKT_BUF_SZ + sizeof(struct RxFD));
1633 PKT_BUF_SZ + sizeof(struct RxFD), PCI_DMA_FROMDEVICE);
1648 rxf->count = cpu_to_le32(PKT_BUF_SZ << 16);
1794 PKT_BUF_SZ + sizeof(struct RxFD),
1858 PKT_BUF_SZ + sizeof(struct RxFD), PCI_DMA_FROMDEVICE);
H A Dvia-rhine.c80 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer.*/ macro
863 PKT_BUF_SZ * TX_RING_SIZE,
893 pci_free_consistent(rp->pdev, PKT_BUF_SZ * TX_RING_SIZE,
908 rp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32);
975 rp->tx_buf[i] = &rp->tx_bufs[i * PKT_BUF_SZ];
1234 if (skb->len > PKT_BUF_SZ) {
H A D82596.c161 #define PKT_BUF_SZ 1536 macro
539 struct sk_buff *skb = dev_alloc_skb(PKT_BUF_SZ);
550 rbd->size = PKT_BUF_SZ;
552 cache_clear(virt_to_phys(skb->data), PKT_BUF_SZ);
808 newskb = dev_alloc_skb(PKT_BUF_SZ);
821 cache_clear(virt_to_phys(newskb->data), PKT_BUF_SZ);
H A Dtyphoon.c55 #define PKT_BUF_SZ 1536 macro
1590 skb = dev_alloc_skb(PKT_BUF_SZ);
1597 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
1653 PKT_BUF_SZ,
1657 PKT_BUF_SZ,
1664 pci_unmap_single(tp->pdev, dma_addr, PKT_BUF_SZ,
1799 pci_unmap_single(tp->pdev, rxb->dma_addr, PKT_BUF_SZ,
1890 xp_cmd.parm1 = cpu_to_le16(PKT_BUF_SZ);
H A D3c59x.c32 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer.*/ macro
1620 iowrite8(PKT_BUF_SZ>>8, ioaddr + TxFreeThreshold); /* Room for a packet. */
1676 vp->rx_ring[i].length = cpu_to_le32(PKT_BUF_SZ | LAST_FRAG);
1677 skb = dev_alloc_skb(PKT_BUF_SZ);
1683 vp->rx_ring[i].addr = cpu_to_le32(pci_map_single(VORTEX_PCI(vp), skb->data, PKT_BUF_SZ, PCI_DMA_FROMDEVICE));
1865 iowrite8(PKT_BUF_SZ>>8, ioaddr + TxFreeThreshold);
2473 pci_dma_sync_single_for_cpu(VORTEX_PCI(vp), dma, PKT_BUF_SZ, PCI_DMA_FROMDEVICE); local
2478 pci_dma_sync_single_for_device(VORTEX_PCI(vp), dma, PKT_BUF_SZ, PCI_DMA_FROMDEVICE); local
2485 pci_unmap_single(VORTEX_PCI(vp), dma, PKT_BUF_SZ, PCI_DMA_FROMDEVICE); local
2509 skb = dev_alloc_skb(PKT_BUF_SZ);
[all...]
H A D3c515.c58 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer. */ macro
816 vp->rx_ring[i].length = PKT_BUF_SZ | 0x80000000;
817 skb = dev_alloc_skb(PKT_BUF_SZ);
830 outb(PKT_BUF_SZ >> 8, ioaddr + TxFreeThreshold); /* Room for a packet. */
1400 skb = dev_alloc_skb(PKT_BUF_SZ);
H A Dvia-velocity.c959 tsize = vptr->options.numtx * PKT_BUF_SZ * vptr->num_txq;
970 memset(vptr->tx_bufs, 0, vptr->options.numtx * PKT_BUF_SZ * vptr->num_txq);
1002 size = vptr->options.numtx * PKT_BUF_SZ * vptr->num_txq;
1162 (j * vptr->options.numtx + i) * PKT_BUF_SZ;
1164 (j * vptr->options.numtx + i) * PKT_BUF_SZ;
1726 vptr->rx_buf_sz = (dev->mtu <= 1504 ? PKT_BUF_SZ : dev->mtu + 32);
H A Ddeclance.c162 #define PKT_BUF_SZ 1536 macro
163 #define RX_BUFF_SIZE PKT_BUF_SZ
164 #define TX_BUFF_SIZE PKT_BUF_SZ
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/tulip/
H A Dinterrupt.c76 skb = tp->rx_buffers[entry].skb = dev_alloc_skb(PKT_BUF_SZ);
80 mapping = pci_map_single(tp->pdev, skb->data, PKT_BUF_SZ,
228 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
451 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
H A Dxircom_tulip_cb.c80 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer.*/ macro
867 tp->rx_ring[i].length = PKT_BUF_SZ;
872 tp->rx_ring[i-1].length = PKT_BUF_SZ | Rx1RingWrap;
879 struct sk_buff *skb = dev_alloc_skb(PKT_BUF_SZ);
896 tp->tx_aligned_skbuff[i] = dev_alloc_skb(PKT_BUF_SZ);
1270 skb = tp->rx_skbuff[entry] = dev_alloc_skb(PKT_BUF_SZ);
H A Dtulip.h271 #define PKT_BUF_SZ 1536 /* Size of each temporary Rx buffer. */ macro
H A Dtulip_core.c597 tp->rx_ring[i].length = cpu_to_le32(PKT_BUF_SZ);
603 tp->rx_ring[i-1].length = cpu_to_le32(PKT_BUF_SZ | DESC_RING_WRAP);
612 struct sk_buff *skb = dev_alloc_skb(PKT_BUF_SZ);
617 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
792 pci_unmap_single(tp->pdev, mapping, PKT_BUF_SZ,
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/tokenring/
H A D3c359.h230 #define PKT_BUF_SZ 4096 /* Default packet size */ macro
H A Dlanstreamer.h239 #define PKT_BUF_SZ 4096 /* Default packet size */ macro
H A Dolympic.h212 #define PKT_BUF_SZ 4096 /* Default packet size */ macro
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/wan/lmc/
H A Dlmc_var.h425 #define PKT_BUF_SZ 1542 /* was 1536 */ macro

Completed in 320 milliseconds

12