Lines Matching refs:end
136 static bool host_stage2_force_pte_cb(u64 addr, u64 end, enum kvm_pgtable_prot prot);
164 static bool guest_stage2_force_pte_cb(u64 addr, u64 end,
345 u64 end;
352 phys_addr_t end;
355 range->end = ULONG_MAX;
361 end = reg->base + reg->size;
364 range->end = reg->base;
365 } else if (addr >= end) {
367 range->start = end;
370 range->end = end;
397 return range->start <= addr && addr < range->end;
400 static bool range_is_memory(u64 start, u64 end)
407 return is_in_mem_range(end - 1, &r);
410 static inline int __host_stage2_idmap(u64 start, u64 end,
413 return kvm_pgtable_stage2_map(&host_mmu.pgt, start, end - start, start,
439 return parent->start <= child->start && child->end <= parent->end;
463 cur.end = cur.start + granule;
486 static bool host_stage2_force_pte_cb(u64 addr, u64 end, enum kvm_pgtable_prot prot)
502 if (range_is_memory(addr, end))
522 ret = host_stage2_idmap_locked(range.start, range.end - range.start, prot);
806 void *start = (void *)addr, *end = start + (tx->nr_pages * PAGE_SIZE);
810 return pkvm_create_mappings_locked(start, end, prot);
824 void *start = (void *)addr, *end = start + (tx->nr_pages * PAGE_SIZE);
827 return pkvm_create_mappings_locked(start, end, prot);
854 * end when we forward the FFA call.
1225 u64 end = PAGE_ALIGN((u64)to);
1226 u64 size = end - start;
1242 for (cur = start; cur < end; cur += PAGE_SIZE)
1255 u64 end = PAGE_ALIGN((u64)to);
1260 for (cur = start; cur < end; cur += PAGE_SIZE)