• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.10.1/postfix-255/postfix/src/util/

Lines Matching defs:titan_ge_eth

242 	titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
248 schedule_work(&titan_ge_eth->tx_timeout_task);
254 static void titan_ge_update_afx(titan_ge_port_info * titan_ge_eth)
256 int port = titan_ge_eth->port_num;
261 memcpy(p_addr, titan_ge_eth->port_mac_addr, 6);
309 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
310 int port = titan_ge_eth->port_num;
328 titan_ge_port_reset(titan_ge_eth->port_num);
329 titan_ge_port_start(netdev, titan_ge_eth);
338 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
344 spin_lock_irqsave(&titan_ge_eth->lock, flags);
358 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
363 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
375 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
376 unsigned int port_num = titan_ge_eth->port_num;
442 titan_ge_free_tx_queue(titan_ge_eth);
533 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
534 unsigned int port_num = titan_ge_eth->port_num;
570 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
571 unsigned int port_num = titan_ge_eth->port_num;
586 spin_lock_irq(&(titan_ge_eth->lock));
589 spin_unlock_irq(&(titan_ge_eth->lock));
595 spin_unlock_irq(&(titan_ge_eth->lock));
658 titan_ge_port_info * titan_ge_eth)
662 titan_ge_port_reset(titan_ge_eth->port_num);
1079 static void titan_ge_tx_queue(titan_ge_port_info * titan_ge_eth,
1082 struct device *device = &titan_ge_device[titan_ge_eth->port_num]->dev;
1083 unsigned int curr_desc = titan_ge_eth->tx_curr_desc_q;
1085 int port_num = titan_ge_eth->port_num;
1087 tx_curr = &(titan_ge_eth->tx_desc_area[curr_desc]);
1092 titan_ge_eth->tx_skb[curr_desc] = (struct sk_buff *) skb;
1102 titan_ge_eth->tx_curr_desc_q = (curr_desc + 1) % TITAN_GE_TX_QUEUE;
1106 &titan_ge_eth->tx_desc_area[titan_ge_eth->tx_curr_desc_q]);
1114 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1115 unsigned int port_num = titan_ge_eth->port_num;
1135 memcpy(titan_ge_eth->port_mac_addr, netdev->dev_addr, 6);
1138 titan_port_init(netdev, titan_ge_eth);
1140 titan_ge_update_afx(titan_ge_eth);
1143 titan_ge_eth->tx_ring_skbs = 0;
1144 titan_ge_eth->tx_ring_size = TITAN_GE_TX_QUEUE;
1147 titan_ge_eth->tx_desc_area = (titan_ge_tx_desc *)
1149 titan_ge_eth->tx_dma = TITAN_SRAM_BASE + TITAN_TX_RING_BYTES * port_num;
1151 if (!titan_ge_eth->tx_desc_area) {
1158 memset(titan_ge_eth->tx_desc_area, 0, titan_ge_eth->tx_desc_area_size);
1161 titan_ge_init_tx_desc_ring(titan_ge_eth,
1162 titan_ge_eth->tx_ring_size,
1163 (unsigned long) titan_ge_eth->tx_desc_area,
1164 (unsigned long) titan_ge_eth->tx_dma);
1167 titan_ge_eth->rx_ring_size = TITAN_GE_RX_QUEUE;
1168 titan_ge_eth->rx_ring_skbs = 0;
1170 titan_ge_eth->rx_desc_area =
1173 titan_ge_eth->rx_dma = TITAN_SRAM_BASE + 0x1000 + TITAN_RX_RING_BYTES * port_num;
1175 if (!titan_ge_eth->rx_desc_area) {
1182 dma_free_coherent(device, titan_ge_eth->tx_desc_area_size,
1183 (void *) titan_ge_eth->tx_desc_area,
1184 titan_ge_eth->tx_dma);
1189 memset(titan_ge_eth->rx_desc_area, 0, titan_ge_eth->rx_desc_area_size);
1194 (titan_ge_eth, titan_ge_eth->rx_ring_size, TITAN_GE_JUMBO_BUFSIZE,
1195 (unsigned long) titan_ge_eth->rx_desc_area, 0,
1196 (unsigned long) titan_ge_eth->rx_dma)) == 0)
1199 (titan_ge_eth, titan_ge_eth->rx_ring_size, TITAN_GE_STD_BUFSIZE,
1200 (unsigned long) titan_ge_eth->rx_desc_area, 0,
1201 (unsigned long) titan_ge_eth->rx_dma)) == 0)
1206 titan_ge_port_start(netdev, titan_ge_eth);
1216 titan_ge_eth->tx_int_coal =
1246 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1251 stats = &titan_ge_eth->stats;
1252 spin_lock_irqsave(&titan_ge_eth->lock, flags);
1254 if ((TITAN_GE_TX_QUEUE - titan_ge_eth->tx_ring_skbs) <=
1257 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
1262 titan_ge_tx_queue(titan_ge_eth, skb);
1263 titan_ge_eth->tx_ring_skbs++;
1265 if (TITAN_GE_TX_QUEUE <= (titan_ge_eth->tx_ring_skbs + 4)) {
1266 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
1267 titan_ge_free_tx_queue(titan_ge_eth);
1268 spin_lock_irqsave(&titan_ge_eth->lock, flags);
1274 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
1325 static int titan_ge_free_tx_queue(titan_ge_port_info *titan_ge_eth)
1330 spin_lock_irqsave(&(titan_ge_eth->lock), flags);
1332 while (titan_ge_return_tx_desc(titan_ge_eth, titan_ge_eth->port_num) == 0)
1333 if (titan_ge_eth->tx_ring_skbs != 1)
1334 titan_ge_eth->tx_ring_skbs--;
1336 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
1354 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1355 unsigned int port_num = titan_ge_eth->port_num;
1362 stats = &titan_ge_eth->stats;
1365 && (titan_ge_rx(netdev, port_num, titan_ge_eth, &packet) == TITAN_OK)) {
1368 titan_ge_eth->rx_ring_skbs--;
1370 if (--titan_ge_eth->rx_work_limit < 0)
1403 if (titan_ge_eth->rx_threshold > RX_THRESHOLD) {
1404 ack = titan_ge_rx_task(netdev, titan_ge_eth);
1406 titan_ge_eth->rx_threshold = 0;
1408 titan_ge_eth->rx_threshold++;
1410 if (titan_ge_eth->tx_threshold > TX_THRESHOLD) {
1411 titan_ge_eth->tx_threshold = 0;
1412 titan_ge_free_tx_queue(titan_ge_eth);
1415 titan_ge_eth->tx_threshold++;
1426 titan_ge_port_info *titan_ge_eth,
1449 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1450 int port_num = titan_ge_eth->port_num;
1454 titan_ge_eth->rx_work_limit = *budget;
1455 if (titan_ge_eth->rx_work_limit > netdev->quota)
1456 titan_ge_eth->rx_work_limit = netdev->quota;
1460 titan_ge_free_tx_queue(titan_ge_eth);
1473 if (titan_ge_eth->rx_work_limit < 0)
1510 spin_lock_irqsave(&titan_ge_eth->lock, flags);
1516 titan_ge_enable_int(port_num, titan_ge_eth, netdev);
1518 spin_unlock_irqrestore(&titan_ge_eth->lock, flags);
1528 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1530 spin_lock_irq(&(titan_ge_eth->lock));
1533 spin_unlock_irq(&titan_ge_eth->lock);
1543 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1544 unsigned int port_num = titan_ge_eth->port_num;
1563 (titan_ge_eth->tx_ring_skbs) && (curr < TITAN_GE_TX_QUEUE);
1565 if (titan_ge_eth->tx_skb[curr]) {
1566 dev_kfree_skb(titan_ge_eth->tx_skb[curr]);
1567 titan_ge_eth->tx_ring_skbs--;
1571 if (titan_ge_eth->tx_ring_skbs != 0)
1575 titan_ge_eth->tx_ring_skbs);
1579 titan_ge_eth->tx_desc_area_size,
1580 (void *) titan_ge_eth->tx_desc_area,
1581 titan_ge_eth->tx_dma);
1590 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1591 unsigned int port_num = titan_ge_eth->port_num;
1610 titan_ge_eth->rx_ring_skbs && (curr < TITAN_GE_RX_QUEUE);
1612 if (titan_ge_eth->rx_skb[curr]) {
1613 dev_kfree_skb(titan_ge_eth->rx_skb[curr]);
1614 titan_ge_eth->rx_ring_skbs--;
1618 if (titan_ge_eth->rx_ring_skbs != 0)
1622 titan_ge_eth->rx_ring_skbs);
1626 titan_ge_eth->rx_desc_area_size,
1627 (void *) titan_ge_eth->rx_desc_area,
1628 titan_ge_eth->rx_dma);
1637 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1641 titan_ge_port_reset(titan_ge_eth->port_num);
1657 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1658 unsigned int port_num = titan_ge_eth->port_num;
1661 memcpy(titan_ge_eth->port_mac_addr, netdev->dev_addr, 6);
1665 titan_ge_update_afx(titan_ge_eth);
1709 titan_ge_port_info *titan_ge_eth = netdev_priv(netdev);
1711 return &titan_ge_eth->stats;
1808 titan_ge_port_info *titan_ge_eth;
1840 titan_ge_eth = netdev_priv(netdev);
1843 titan_ge_eth->port_num = port;
1846 INIT_WORK(&titan_ge_eth->tx_timeout_task,
1849 spin_lock_init(&titan_ge_eth->lock);
1909 static int titan_ge_return_tx_desc(titan_ge_port_info * titan_ge_eth, int port)
1914 tx_desc_used = titan_ge_eth->tx_used_desc_q;
1917 if (tx_desc_used == titan_ge_eth->tx_curr_desc_q)
1921 skb = titan_ge_eth->tx_skb[tx_desc_used];
1925 titan_ge_eth->tx_skb[tx_desc_used] = NULL;
1926 titan_ge_eth->tx_used_desc_q =