Lines Matching defs:cur_seg
271 struct xhci_segment *cur_seg;
285 for (cur_seg = enq_seg->next; cur_seg != enq_seg;
286 cur_seg = cur_seg->next)
388 struct xhci_segment *cur_seg = start_seg;
391 while (cur_seg->trbs > trb ||
392 &cur_seg->trbs[TRBS_PER_SEGMENT - 1] < trb) {
393 generic_trb = &cur_seg->trbs[TRBS_PER_SEGMENT - 1].generic;
398 cur_seg = cur_seg->next;
399 if (cur_seg == start_seg)
403 return cur_seg;
523 struct xhci_segment *cur_seg;
526 for (cur_seg = cur_td->start_seg, cur_trb = cur_td->first_trb;
528 next_trb(xhci, ep_ring, &cur_seg, &cur_trb)) {
539 (unsigned long long)xhci_trb_virt_to_dma(cur_seg, cur_trb),
540 cur_seg,
541 (unsigned long long)cur_seg->dma);
552 (unsigned long long)xhci_trb_virt_to_dma(cur_seg, cur_trb),
553 cur_seg,
554 (unsigned long long)cur_seg->dma);
1190 struct xhci_segment *cur_seg;
1193 cur_seg = start_seg;
1199 end_seg_dma = xhci_trb_virt_to_dma(cur_seg,
1200 &cur_seg->trbs[TRBS_PER_SEGMENT - 1]);
1202 end_trb_dma = xhci_trb_virt_to_dma(cur_seg, end_trb);
1208 return cur_seg;
1215 (suspect_dma >= cur_seg->dma &&
1217 return cur_seg;
1223 return cur_seg;
1225 cur_seg = cur_seg->next;
1226 start_dma = xhci_trb_virt_to_dma(cur_seg, &cur_seg->trbs[0]);
1227 } while (cur_seg != start_seg);
1512 struct xhci_segment *cur_seg;
1577 cur_seg = ep_ring->deq_seg; cur_trb != event_trb;
1578 next_trb(xhci, ep_ring, &cur_seg, &cur_trb)) {
1610 struct xhci_segment *cur_seg;
1690 for (cur_trb = ep_ring->dequeue, cur_seg = ep_ring->deq_seg;
1692 next_trb(xhci, ep_ring, &cur_seg, &cur_trb)) {