/linux-master/drivers/net/ethernet/sun/ |
H A D | sungem.c | 653 last &= (TX_RING_SIZE - 1); 973 if (!(entry & ((TX_RING_SIZE>>1)-1))) 1580 for (i = 0; i < TX_RING_SIZE; i++) { 1589 int ent = i & (TX_RING_SIZE - 1); 1643 for (i = 0; i < TX_RING_SIZE; i++) {
|
H A D | sunhme.c | 1176 for (i = 0; i < TX_RING_SIZE; i++) { 1249 for (i = 0; i < TX_RING_SIZE; i++) 1443 HMD("tx rsize=%d oreg[%08x]\n", (int)TX_RING_SIZE, 1445 hme_write32(hp, etxregs + ETX_RSIZE, (TX_RING_SIZE >> ETX_RSIZE_SHIFT) - 1); 1719 last &= (TX_RING_SIZE - 1);
|
H A D | sunbmac.c | 204 for (i = 0; i < TX_RING_SIZE; i++) { 249 for (i = 0; i < TX_RING_SIZE; i++)
|
H A D | sunqe.c | 588 txbuf = &qbufs->tx_buf[entry & (TX_RING_SIZE - 1)][0]; 590 qebuf_offset(tx_buf, (entry & (TX_RING_SIZE - 1)));
|
/linux-master/drivers/net/ethernet/renesas/ |
H A D | sh_eth.c | 3277 mdp->num_tx_ring = TX_RING_SIZE;
|
H A D | rswitch.h | 28 #define TX_RING_SIZE 1024 macro 30 #define TS_RING_SIZE (TX_RING_SIZE * RSWITCH_NUM_PORTS)
|
H A D | rswitch.c | 587 err = rswitch_gwca_queue_alloc(ndev, priv, rdev->tx_queue, true, TX_RING_SIZE);
|
/linux-master/drivers/net/ethernet/pasemi/ |
H A D | pasemi_mac.c | 466 ring->size = TX_RING_SIZE; 467 ring->ring_info = kcalloc(TX_RING_SIZE, 474 if (pasemi_dma_alloc_ring(&ring->chan, TX_RING_SIZE)) 480 val |= PAS_DMA_TXCHAN_BASEU_SIZ(TX_RING_SIZE >> 3); 522 limit += TX_RING_SIZE; 525 info = &txring->ring_info[(i+1) & (TX_RING_SIZE-1)]; 530 (TX_RING_SIZE-1)].dma; 840 ring_limit += TX_RING_SIZE; 888 txring->next_to_clean = i & (TX_RING_SIZE-1); 1522 txring->next_to_fill = (fill + nfrags + 1) & (TX_RING_SIZE [all...] |
/linux-master/drivers/net/ethernet/natsemi/ |
H A D | natsemi.c | 102 #define TX_RING_SIZE 16 macro 174 the list. The ring sizes are set at compile time by RX/TX_RING_SIZE. 547 struct sk_buff *tx_skbuff[TX_RING_SIZE]; 548 dma_addr_t tx_dma[TX_RING_SIZE]; 1865 for (i = 0; i < TX_RING_SIZE; i++) { 1917 sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE), 1975 for (i = 0; i < TX_RING_SIZE; i++) { 1979 *((i+1)%TX_RING_SIZE+RX_RING_SIZE)); 2011 for (i = 0; i < TX_RING_SIZE; i++) { 2052 sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE), [all...] |
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 3345 txq->bd.ring_size = TX_RING_SIZE;
|
H A D | fec.h | 359 #define TX_RING_SIZE 1024 /* Must be power of two */ macro 561 unsigned char *tx_bounce[TX_RING_SIZE]; 562 struct fec_tx_buffer tx_buf[TX_RING_SIZE];
|
/linux-master/drivers/net/ethernet/dlink/ |
H A D | sundance.c | 65 #define TX_RING_SIZE 32 macro 66 #define TX_QUEUE_LEN (TX_RING_SIZE - 1) /* Limit ring entries actually used. */ 69 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct netdev_desc) 128 the list. The ring sizes are set at compile time by RX/TX_RING_SIZE. 366 struct sk_buff* tx_skbuff[TX_RING_SIZE]; 977 for (i=0; i<TX_RING_SIZE; i++) { 990 np->cur_tx, np->cur_tx % TX_RING_SIZE, 991 np->dirty_tx, np->dirty_tx % TX_RING_SIZE); 1055 for (i = 0; i < TX_RING_SIZE; i++) { 1064 unsigned head = np->cur_task % TX_RING_SIZE; [all...] |
H A D | dl2k.c | 210 else if (tx_coalesce > TX_RING_SIZE-1) 211 tx_coalesce = TX_RING_SIZE - 1; 450 for (i = 0; i < TX_RING_SIZE; i++) { 471 for (i = 0; i < TX_RING_SIZE; i++) 488 for (i = 0; i < TX_RING_SIZE; i++) { 491 ((i + 1) % TX_RING_SIZE) * 717 entry = np->cur_tx % TX_RING_SIZE; 753 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; 754 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE [all...] |
/linux-master/drivers/net/ethernet/3com/ |
H A D | 3c515.c | 48 #define TX_RING_SIZE 16 macro 299 struct boom_tx_desc tx_ring[TX_RING_SIZE]; 302 struct sk_buff *tx_skbuff[TX_RING_SIZE]; 832 for (i = 0; i < TX_RING_SIZE; i++) 970 for (i = 0; i < TX_RING_SIZE; i++) { 1000 int entry = vp->cur_tx % TX_RING_SIZE; 1008 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; 1038 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) 1163 int entry = dirty_tx % TX_RING_SIZE; 1174 if (lp->tx_full && (lp->cur_tx - dirty_tx <= TX_RING_SIZE [all...] |
/linux-master/drivers/net/ethernet/smsc/ |
H A D | smsc9420.c | 498 for (i = 0; i < TX_RING_SIZE; i++) { 921 pd->tx_ring_tail = (pd->tx_ring_tail + 1) % TX_RING_SIZE; 933 (((pd->tx_ring_head + 2) % TX_RING_SIZE) == pd->tx_ring_tail); 960 if (unlikely(index == (TX_RING_SIZE - 1))) 968 pd->tx_ring_head = (pd->tx_ring_head + 1) % TX_RING_SIZE; 1182 pd->tx_buffers = kmalloc_array(TX_RING_SIZE, 1189 for (i = 0; i < TX_RING_SIZE; i++) { 1197 pd->tx_ring[TX_RING_SIZE - 1].length = TDES1_TER_; 1550 sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE), 1608 sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE), [all...] |
/linux-master/drivers/net/ethernet/via/ |
H A D | via-rhine.c | 77 #define TX_RING_SIZE 64 macro 78 #define TX_QUEUE_LEN (TX_RING_SIZE - 6) /* Limit ring entries actually used. */ 156 the list. The ring sizes are set at compile time by RX/TX_RING_SIZE. 446 struct sk_buff *tx_skbuff[TX_RING_SIZE]; 447 dma_addr_t tx_skbuff_dma[TX_RING_SIZE]; 450 unsigned char *tx_buf[TX_RING_SIZE]; 1147 TX_RING_SIZE * sizeof(struct tx_desc), 1156 PKT_BUF_SZ * TX_RING_SIZE, 1162 TX_RING_SIZE * sizeof(struct tx_desc), 1183 TX_RING_SIZE * sizeo [all...] |
/linux-master/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 67 #define TX_RING_SIZE (TX_RING_ENTRIES * sizeof(struct ioc3_etxd)) macro 894 ip->tx_ring = dma_alloc_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, 958 dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring, 971 dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring, ip->txr_dma);
|
/linux-master/drivers/net/ethernet/freescale/fs_enet/ |
H A D | fs_enet-main.c | 62 #define TX_RING_SIZE 64 macro 95 int tx_left = TX_RING_SIZE; 934 fpi->tx_ring = TX_RING_SIZE;
|
/linux-master/drivers/net/ethernet/dec/tulip/ |
H A D | tulip.h | 326 #define TX_RING_SIZE 32 macro 409 struct ring_info tx_buffers[TX_RING_SIZE];
|
/linux-master/drivers/net/ethernet/amd/ |
H A D | sunlance.c | 176 #define TX_RING_SIZE (1 << (LANCE_LOG_TX_BUFFERS)) macro 177 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1) 223 struct lance_tx_desc btx_ring[TX_RING_SIZE]; 225 u8 tx_buf [TX_RING_SIZE][TX_BUFF_SIZE]; 340 for (i = 0; i < TX_RING_SIZE; i++) { 396 for (i = 0; i < TX_RING_SIZE; i++) {
|
H A D | atarilance.c | 111 #define TX_RING_SIZE (1 << TX_LOG_RING_SIZE) macro 113 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1) 157 struct lance_tx_head tx_head[TX_RING_SIZE]; 704 for( i = 0; i < TX_RING_SIZE; i++ ) { 755 for( i = 0 ; i < TX_RING_SIZE; i++ ) 829 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { 830 lp->cur_tx -= TX_RING_SIZE; 831 lp->dirty_tx -= TX_RING_SIZE; 920 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { [all...] |
H A D | lance.c | 194 #define TX_RING_SIZE (1 << (LANCE_LOG_TX_BUFFERS)) macro 195 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1) 239 struct lance_tx_head tx_ring[TX_RING_SIZE]; 243 struct sk_buff* tx_skbuff[TX_RING_SIZE]; 564 lp->tx_bounce_buffs = kmalloc_array(TX_RING_SIZE, PKT_BUF_SZ, 855 for (i = 0; i < TX_RING_SIZE; i++) { 892 for (i = 0; i < TX_RING_SIZE; i++) { 941 for (i = 0; i < TX_RING_SIZE; i++) 1016 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) 1099 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { [all...] |
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede.h | 596 #define TX_RING_SIZE ((u16)BIT(TX_RING_SIZE_POW)) macro 597 #define NUM_TX_BDS_MAX (TX_RING_SIZE - 1)
|
/linux-master/drivers/net/ethernet/i825xx/ |
H A D | lib82596.c | 270 #define TX_RING_SIZE 32 macro 309 struct tx_cmd tx_cmds[TX_RING_SIZE] __attribute__((aligned(32))); 310 struct i596_tbd tbds[TX_RING_SIZE] __attribute__((aligned(32))); 361 static int max_cmd_backlog = TX_RING_SIZE-1; 1026 if (++lp->next_tx_cmd == TX_RING_SIZE)
|
H A D | 82596.c | 286 #define TX_RING_SIZE 64 macro 333 struct tx_cmd tx_cmds[TX_RING_SIZE]; 334 struct i596_tbd tbds[TX_RING_SIZE]; 371 static int max_cmd_backlog = TX_RING_SIZE-1; 1076 if (++lp->next_tx_cmd == TX_RING_SIZE)
|