Lines Matching refs:qp
489 * Check if the folio's nid is in qp->nmask.
491 * If MPOL_MF_INVERT is set in qp->flags, check if the nid is
492 * in the invert of qp->nmask.
495 struct queue_pages *qp)
498 unsigned long flags = qp->flags;
500 return node_isset(nid, *qp->nmask) == !(flags & MPOL_MF_INVERT);
506 struct queue_pages *qp = walk->private;
509 qp->nr_failed++;
517 if (!queue_folio_required(folio, qp))
519 if (!(qp->flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) ||
521 !migrate_folio_add(folio, qp->pagelist, qp->flags))
522 qp->nr_failed++;
540 struct queue_pages *qp = walk->private;
541 unsigned long flags = qp->flags;
564 qp->nr_failed++;
576 if (!queue_folio_required(folio, qp))
595 if (folio == qp->large)
597 qp->large = folio;
601 !migrate_folio_add(folio, qp->pagelist, flags)) {
602 qp->nr_failed++;
610 if (qp->nr_failed && strictly_unmovable(flags))
620 struct queue_pages *qp = walk->private;
621 unsigned long flags = qp->flags;
630 qp->nr_failed++;
634 if (!queue_folio_required(folio, qp))
638 qp->nr_failed++;
650 if (!isolate_hugetlb(folio, qp->pagelist))
651 qp->nr_failed++;
654 if (qp->nr_failed && strictly_unmovable(flags))
692 struct queue_pages *qp = walk->private;
693 unsigned long flags = qp->flags;
698 if (!qp->first) {
699 qp->first = vma;
701 (qp->start < vma->vm_start))
707 ((vma->vm_end < qp->end) &&
763 struct queue_pages qp = {
774 err = walk_page_range(mm, start, end, ops, &qp);
776 if (!qp.first)
780 return err ? : qp.nr_failed;