Lines Matching refs:walker

63 	struct kvm_pgtable_walker	*walker;
182 struct kvm_pgtable_walker *walker = data->walker;
186 return walker->cb(ctx, visit);
189 static bool kvm_pgtable_walk_continue(const struct kvm_pgtable_walker *walker,
203 return !(walker->flags & KVM_PGTABLE_WALK_HANDLE_FAULT);
215 enum kvm_pgtable_walk_flags flags = data->walker->flags;
216 kvm_pte_t *ptep = kvm_dereference_pteref(data->walker, pteref);
220 .arg = data->walker->arg,
244 * Reload the page table after invoking the walker callback for leaf
245 * entries or after pre-order traversal, to allow the walker to descend
253 if (!kvm_pgtable_walk_continue(data->walker, ret))
264 if (!kvm_pgtable_walk_continue(data->walker, ret))
271 if (kvm_pgtable_walk_continue(data->walker, ret))
325 struct kvm_pgtable_walker *walker)
331 .walker = walker,
335 r = kvm_pgtable_walk_begin(walker);
340 kvm_pgtable_walk_end(walker);
365 struct kvm_pgtable_walker walker = {
373 PAGE_SIZE, &walker);
496 struct kvm_pgtable_walker walker = {
506 ret = kvm_pgtable_walk(pgt, addr, size, &walker);
555 struct kvm_pgtable_walker walker = {
564 kvm_pgtable_walk(pgt, addr, size, &walker);
609 struct kvm_pgtable_walker walker = {
614 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker));
615 pgt->mm_ops->put_page(kvm_dereference_pteref(&walker, pgt->pgd));
831 * Should never occur if this walker has exclusive access to the
883 * pointless when the unmap walker needs to perform CMOs.
932 * walker using an invalid PA. Avoid offsetting an already invalid PA,
1090 struct kvm_pgtable_walker walker = {
1105 ret = kvm_pgtable_walk(pgt, addr, size, &walker);
1121 struct kvm_pgtable_walker walker = {
1131 ret = kvm_pgtable_walk(pgt, addr, size, &walker);
1181 struct kvm_pgtable_walker walker = {
1187 ret = kvm_pgtable_walk(pgt, addr, size, &walker);
1250 struct kvm_pgtable_walker walker = {
1256 ret = kvm_pgtable_walk(pgt, addr, size, &walker);
1330 struct kvm_pgtable_walker walker = {
1336 WARN_ON(kvm_pgtable_walk(pgt, addr, size, &walker));
1384 struct kvm_pgtable_walker walker = {
1393 return kvm_pgtable_walk(pgt, addr, size, &walker);
1407 struct kvm_pgtable_walker walker = {
1420 .walker = &walker,
1542 struct kvm_pgtable_walker walker = {
1548 return kvm_pgtable_walk(pgt, addr, size, &walker);
1574 /* Ensure zeroed PGD pages are visible to the hardware walker */
1607 struct kvm_pgtable_walker walker = {
1613 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker));
1615 pgt->mm_ops->free_pages_exact(kvm_dereference_pteref(&walker, pgt->pgd), pgd_sz);
1622 struct kvm_pgtable_walker walker = {
1628 .walker = &walker,