Lines Matching refs:ring_idx
529 vmxnet3_rq_alloc_rx_buf(struct vmxnet3_rx_queue *rq, u32 ring_idx,
533 struct vmxnet3_rx_buf_info *rbi_base = rq->buf_info[ring_idx];
534 struct vmxnet3_cmd_ring *ring = &rq->rx_ring[ring_idx];
589 rq->uncommitted[ring_idx] += num_allocated;
594 ring->next2comp, rq->uncommitted[ring_idx]);
1100 u32 idx, ring_idx;
1111 ring_idx = rcd->rqID == rq->qid ? 0 : 1;
1112 vmxnet3_getRxDesc(rxd, &rq->rx_ring[ring_idx].base[idx].rxd,
1114 rbi = rq->buf_info[ring_idx] + idx;
1136 ring_idx, idx);
1192 rq->rx_ring[ring_idx].next2comp = idx;
1193 VMXNET3_INC_RING_IDX_ONLY(rq->rx_ring[ring_idx].next2comp,
1194 rq->rx_ring[ring_idx].size);
1198 ring_idx);
1200 ring_idx, adapter))) {
1201 vmxnet3_rq_alloc_rx_buf(rq, ring_idx, num_to_alloc,
1207 rxprod_reg[ring_idx] + rq->qid * 8,
1208 rq->rx_ring[ring_idx].next2fill);
1209 rq->uncommitted[ring_idx] = 0;
1226 u32 i, ring_idx;
1229 for (ring_idx = 0; ring_idx < 2; ring_idx++) {
1230 for (i = 0; i < rq->rx_ring[ring_idx].size; i++) {
1235 &rq->rx_ring[ring_idx].base[i].rxd, &rxDesc);
1238 rq->buf_info[ring_idx][i].skb) {
1241 dev_kfree_skb(rq->buf_info[ring_idx][i].skb);
1242 rq->buf_info[ring_idx][i].skb = NULL;
1244 rq->buf_info[ring_idx][i].page) {
1247 put_page(rq->buf_info[ring_idx][i].page);
1248 rq->buf_info[ring_idx][i].page = NULL;
1252 rq->rx_ring[ring_idx].gen = VMXNET3_INIT_GEN;
1253 rq->rx_ring[ring_idx].next2fill =
1254 rq->rx_ring[ring_idx].next2comp = 0;
1255 rq->uncommitted[ring_idx] = 0;