/freebsd-10-stable/sys/ofed/drivers/infiniband/ulp/sdp/ |
H A D | sdp_tx.c | 52 if (!callout_pending(&ssk->tx_ring.timer)) 53 callout_reset(&ssk->tx_ring.timer, SDP_TX_POLL_TIMEOUT, 57 if (force || (++ssk->tx_ring.poll_cnt & (SDP_TX_POLL_MODER - 1)) == 0) 85 mseq = ring_head(ssk->tx_ring); 121 tx_req = &ssk->tx_ring.buffer[mseq & (SDP_TX_SIZE - 1)]; 157 atomic_inc(&ssk->tx_ring.head); 158 atomic_dec(&ssk->tx_ring.credits); 170 struct sdp_tx_ring *tx_ring = &ssk->tx_ring; local 172 if (unlikely(mseq != ring_tail(*tx_ring))) { [all...] |
H A D | sdp.h | 149 #define tx_ring_posted(ssk) (ring_posted(ssk->tx_ring) + \ 150 (ssk->tx_ring.rdma_inflight ? ssk->tx_ring.rdma_inflight->busy : 0)) 152 #define tx_ring_posted(ssk) ring_posted(ssk->tx_ring) 322 #define tx_credits(ssk) (atomic_read(&ssk->tx_ring.credits)) 403 struct sdp_tx_ring tx_ring; member in struct:sdp_sock 500 ib_req_notify_cq(ssk->tx_ring.cq, IB_CQ_NEXT_COMP); 505 * - free slots in tx_ring (not including SDP_MIN_TX_CREDITS
|
H A D | sdp_cma.c | 112 qp_init_attr.send_cq = ssk->tx_ring.cq; 173 atomic_set(&ssk->tx_ring.credits, ssk->max_bufs); 205 atomic_set(&ssk->tx_ring.credits, ssk->max_bufs);
|
/freebsd-10-stable/sys/dev/ixgbe/ |
H A D | ixgbe_fdir.h | 48 void ixgbe_atr(struct tx_ring *, struct mbuf *);
|
H A D | ixv_fdir.h | 48 void ixgbe_atr(struct tx_ring *, struct mbuf *);
|
H A D | ix_txrx.c | 73 static void ixgbe_setup_transmit_ring(struct tx_ring *); 74 static void ixgbe_free_transmit_buffers(struct tx_ring *); 79 static int ixgbe_xmit(struct tx_ring *, struct mbuf **); 80 static int ixgbe_tx_ctx_setup(struct tx_ring *, 82 static int ixgbe_tso_setup(struct tx_ring *, 103 ixgbe_legacy_start_locked(struct ifnet *ifp, struct tx_ring *txr) 145 struct tx_ring *txr = adapter->tx_rings; 164 struct tx_ring *txr; 214 ixgbe_mq_start_locked(struct ifnet *ifp, struct tx_ring *txr) 270 struct tx_ring *tx [all...] |
H A D | ixv_txrx.c | 62 static void ixgbe_setup_transmit_ring(struct tx_ring *); 63 static void ixgbe_free_transmit_buffers(struct tx_ring *); 68 static int ixgbe_xmit(struct tx_ring *, struct mbuf **); 69 static int ixgbe_tx_ctx_setup(struct tx_ring *, 71 static int ixgbe_tso_setup(struct tx_ring *, 92 ixv_legacy_start_locked(struct ifnet *ifp, struct tx_ring *txr) 134 struct tx_ring *txr = adapter->tx_rings; 153 struct tx_ring *txr; 203 ixv_mq_start_locked(struct ifnet *ifp, struct tx_ring *txr) 254 struct tx_ring *tx [all...] |
H A D | ixgbe.h | 313 struct tx_ring *txr; 323 struct tx_ring { struct 494 struct tx_ring *tx_rings; 674 int ixgbe_legacy_start_locked(struct ifnet *, struct tx_ring *); 676 int ixgbe_mq_start_locked(struct ifnet *, struct tx_ring *); 686 void ixgbe_txeof(struct tx_ring *);
|
H A D | ixv.h | 314 struct tx_ring *txr; 324 struct tx_ring { struct 495 struct tx_ring *tx_rings; 675 int ixv_legacy_start_locked(struct ifnet *, struct tx_ring *); 677 int ixv_mq_start_locked(struct ifnet *, struct tx_ring *); 687 void ixv_txeof(struct tx_ring *);
|
H A D | if_fdir.c | 80 ixgbe_atr(struct tx_ring *txr, struct mbuf *mp)
|
/freebsd-10-stable/sys/dev/qlxge/ |
H A D | qls_hw.c | 645 txr_done = ha->tx_ring[txr_idx].txr_done; 646 txr_next = ha->tx_ring[txr_idx].txr_next; 649 ha->tx_ring[txr_idx].txr_free = NUM_TX_DESCRIPTORS; 651 ha->tx_ring[txr_idx].txr_free = txr_done - txr_next; 653 ha->tx_ring[txr_idx].txr_free = NUM_TX_DESCRIPTORS + 657 if (ha->tx_ring[txr_idx].txr_free <= QLA_TX_MIN_FREE) 691 if (ha->tx_ring[txr_idx].txr_free <= (NUM_TX_DESCRIPTORS >> 2)) { 695 ha->tx_ring[txr_idx].txr_free); 700 tx_mac = (q81_tx_mac_t *)&ha->tx_ring[txr_idx].wq_vaddr[txr_next]; 712 ha->tx_ring[txr_id [all...] |
H A D | qls_isr.c | 57 txb = &ha->tx_ring[txr_idx].tx_buf[tx_idx]; 69 ha->tx_ring[txr_idx].txr_done++; 71 if (ha->tx_ring[txr_idx].txr_done == NUM_TX_DESCRIPTORS) 72 ha->tx_ring[txr_idx].txr_done = 0;
|
H A D | qls_os.c | 169 "%s: tx_ring[%d].tx_frames= %p\n", 171 (void *)ha->tx_ring[i].tx_frames); 174 "%s: tx_ring[%d].tx_tso_frames= %p\n", 176 (void *)ha->tx_ring[i].tx_tso_frames); 179 "%s: tx_ring[%d].tx_vlan_frames= %p\n", 181 (void *)ha->tx_ring[i].tx_vlan_frames); 184 "%s: tx_ring[%d].txr_free= 0x%08x\n", 186 ha->tx_ring[i].txr_free); 189 "%s: tx_ring[%d].txr_next= 0x%08x\n", 191 ha->tx_ring[ [all...] |
/freebsd-10-stable/sys/dev/et/ |
H A D | if_et.c | 862 struct et_txdesc_ring *tx_ring; local 881 tx_ring = &sc->sc_tx_ring; 883 &tx_ring->tr_dtag, (uint8_t **)&tx_ring->tr_desc, &tx_ring->tr_dmap, 884 &tx_ring->tr_paddr, "TX ring"); 1041 struct et_txdesc_ring *tx_ring; local 1117 tx_ring = &sc->sc_tx_ring; 1118 et_dma_ring_free(sc, &tx_ring->tr_dtag, (void *)&tx_ring 1376 struct et_txdesc_ring *tx_ring; local 1499 struct et_txdesc_ring *tx_ring; local 1682 struct et_txdesc_ring *tx_ring; local 1810 struct et_txdesc_ring *tx_ring; local 2154 struct et_txdesc_ring *tx_ring; local 2251 struct et_txdesc_ring *tx_ring; local [all...] |
/freebsd-10-stable/sys/ofed/drivers/infiniband/hw/mlx4/ |
H A D | mad.c | 532 tun_mad = (struct mlx4_rcv_tunnel_mad *) (tun_qp->tx_ring[tun_tx_ix].buf.addr); 533 if (tun_qp->tx_ring[tun_tx_ix].ah) 534 ib_destroy_ah(tun_qp->tx_ring[tun_tx_ix].ah); 535 tun_qp->tx_ring[tun_tx_ix].ah = ah; 537 tun_qp->tx_ring[tun_tx_ix].buf.map, 554 tun_qp->tx_ring[tun_tx_ix].buf.map, 558 list.addr = tun_qp->tx_ring[tun_tx_ix].buf.map; 1357 sqp_mad = (struct mlx4_mad_snd_buf *) (sqp->tx_ring[wire_tx_ix].buf.addr); 1358 if (sqp->tx_ring[wire_tx_ix].ah) 1359 ib_destroy_ah(sqp->tx_ring[wire_tx_i [all...] |
/freebsd-10-stable/sys/dev/xen/netback/ |
H A D | netback.c | 164 const netif_tx_back_ring_t *tx_ring, 328 netif_tx_back_ring_t tx_ring; member in union:xnb_ring_config::__anon1891 551 &xnb->ring_configs[XNB_RING_TYPE_TX].back_ring.tx_ring; 734 BACK_RING_INIT(&ring->back_ring.tx_ring, 1428 txb = &xnb->ring_configs[XNB_RING_TYPE_TX].back_ring.tx_ring; 1467 * \param[in] tx_ring Pointer to the Ring that is the input to this function 1472 xnb_ring2pkt(struct xnb_pkt *pkt, const netif_tx_back_ring_t *tx_ring, argument 1492 if (RING_HAS_UNCONSUMED_REQUESTS_2(tx_ring, idx)) { 1493 netif_tx_request_t *tx = RING_GET_REQUEST(tx_ring, idx); 1504 RING_HAS_UNCONSUMED_REQUESTS_2(tx_ring, id [all...] |
/freebsd-10-stable/sys/contrib/octeon-sdk/ |
H A D | cvmx-mgmt-port.c | 95 int tx_write_index; /* Where the next TX will write in the tx_ring and tx_buffers */ 99 cvmx_mgmt_port_ring_entry_t tx_ring[CVMX_MGMT_PORT_NUM_TX_BUFFERS]; member in struct:__anon7313 183 if (oring1.s.osize != CVMX_MGMT_PORT_NUM_TX_BUFFERS || cvmx_mgmt_port_state_ptr[port].tx_ring[0].u64 == 0) 199 if (cvmx_mgmt_port_state_ptr[port].tx_ring[0].u64 == 0) 286 state->tx_ring[i].s.len = CVMX_MGMT_PORT_TX_BUFFER_SIZE; 287 state->tx_ring[i].s.addr = cvmx_ptr_to_phys(state->tx_buffers[i]); 292 oring1.s.obase = cvmx_ptr_to_phys(state->tx_ring)>>3; 562 state->tx_ring[state->tx_write_index].s.len = packet_len; 564 state->tx_ring[state->tx_write_index].s.tstamp = 0; 617 state->tx_ring[stat [all...] |
/freebsd-10-stable/sys/dev/netmap/ |
H A D | if_em_netmap.h | 48 struct tx_ring *txr = adapter->tx_rings; 69 struct tx_ring *txr = adapter->tx_rings; 131 struct tx_ring *txr = &adapter->tx_rings[kring->ring_id];
|
/freebsd-10-stable/sys/dev/rt/ |
H A D | if_rt.c | 334 error = rt_alloc_tx_ring(sc, &sc->tx_ring[i], i); 432 rt_free_tx_ring(sc, &sc->tx_ring[i]); 549 rt_free_tx_ring(sc, &sc->tx_ring[i]); 685 rt_reset_tx_ring(sc, &sc->tx_ring[i]); 690 sc->tx_ring[i].desc_phys_addr); 854 RT_SOFTC_TX_RING_ASSERT_LOCKED(&sc->tx_ring[qid]); 857 ring = &sc->tx_ring[qid]; 996 RT_SOFTC_TX_RING_LOCK(&sc->tx_ring[qid]); 998 if (sc->tx_ring[qid].data_queued >= 1000 RT_SOFTC_TX_RING_UNLOCK(&sc->tx_ring[qi [all...] |
/freebsd-10-stable/sys/dev/e1000/ |
H A D | if_em.c | 243 struct tx_ring *); 247 static void em_start_locked(struct ifnet *, struct tx_ring *); 269 static int em_allocate_transmit_buffers(struct tx_ring *); 271 static void em_free_transmit_buffers(struct tx_ring *); 283 static void em_txeof(struct tx_ring *); 291 static void em_transmit_checksum_setup(struct tx_ring *, struct mbuf *, int, 293 static void em_tso_setup(struct tx_ring *, struct mbuf *, int, struct ip *, 303 static int em_xmit(struct tx_ring *, struct mbuf **); 953 struct tx_ring *txr = adapter->tx_rings; 984 em_start_locked(struct ifnet *ifp, struct tx_ring *tx [all...] |
H A D | if_igb.c | 126 static int igb_mq_start_locked(struct ifnet *, struct tx_ring *); 131 static void igb_start_locked(struct tx_ring *, struct ifnet *ifp); 151 static int igb_allocate_transmit_buffers(struct tx_ring *); 153 static void igb_setup_transmit_ring(struct tx_ring *); 156 static void igb_free_transmit_buffers(struct tx_ring *); 169 static bool igb_txeof(struct tx_ring *); 177 static int igb_tx_ctx_setup(struct tx_ring *, 179 static int igb_tso_setup(struct tx_ring *, 191 static int igb_xmit(struct tx_ring *, struct mbuf **); 789 struct tx_ring *tx [all...] |
H A D | if_igb.h | 345 struct tx_ring *txr; 355 struct tx_ring { struct 508 struct tx_ring *tx_rings;
|
/freebsd-10-stable/sys/ofed/drivers/net/mlx4/ |
H A D | en_netdev.c | 1225 struct mlx4_en_tx_ring *tx_ring; local 1321 tx_ring = priv->tx_ring[i]; 1323 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, 1335 for (j = 0; j < tx_ring->buf_size; j += STAMP_STRIDE) 1336 *((u32 *) (tx_ring->buf + j)) = 0xffffffff; 1397 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[tx_index]); 1491 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[i]); 1497 mlx4_en_free_tx_buf(dev, priv->tx_ring[i]); 1531 ring = priv->tx_ring[ 2641 struct mlx4_en_tx_ring *tx_ring; local [all...] |
H A D | en_port.c | 151 const struct mlx4_en_tx_ring *ring = priv->tx_ring[i]; 231 priv->port_stats.tx_chksum_offload += priv->tx_ring[i]->tx_csum; 232 priv->port_stats.queue_stopped += priv->tx_ring[i]->queue_stopped; 233 priv->port_stats.wake_queue += priv->tx_ring[i]->wake_queue; 234 priv->port_stats.oversized_packets += priv->tx_ring[i]->oversized_packets; 235 priv->port_stats.tso_packets += priv->tx_ring[i]->tso_packets; 236 priv->port_stats.defrag_attempts += priv->tx_ring[i]->defrag_attempts;
|
H A D | en_tx.c | 378 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; 461 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; 474 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; 500 struct mlx4_en_tx_ring *ring = priv->tx_ring[tx_ind]; 672 struct mlx4_en_tx_ring *ring = priv->tx_ring[tx_ind]; 976 ring = priv->tx_ring[tx_ind]; 1026 ring = priv->tx_ring[tx_ind]; 1059 ring = priv->tx_ring[i]; 1088 ring = priv->tx_ring[i];
|