Lines Matching refs:addr

147 	vm_offset_t addr;
150 if (vmem_alloc(kernel_arena, size, M_BESTFIT | M_NOWAIT, &addr))
153 return (addr);
166 kva_free(vm_offset_t addr, vm_size_t size)
170 vmem_free(kernel_arena, addr, size);
216 vm_offset_t addr, i, offset;
224 if (vmem_alloc(vmem, size, M_BESTFIT | flags, &addr))
226 offset = addr - VM_MIN_KERNEL_ADDRESS;
235 kmem_unback(object, addr, i);
236 vmem_free(vmem, addr, size);
245 pmap_enter(kernel_pmap, addr + i, m, prot,
249 return (addr);
266 vm_offset_t addr;
271 addr = kmem_alloc_attr_domain(domain, size, flags, low, high,
273 if (addr != 0)
277 return (addr);
295 vm_offset_t addr, offset, tmp;
303 if (vmem_alloc(vmem, size, flags | M_BESTFIT, &addr))
305 offset = addr - VM_MIN_KERNEL_ADDRESS;
313 vmem_free(vmem, addr, size);
320 tmp = addr;
330 return (addr);
348 vm_offset_t addr;
353 addr = kmem_alloc_contig_domain(domain, size, flags, low, high,
355 if (addr != 0)
359 return (addr);
404 vm_offset_t addr;
412 if (vmem_alloc(arena, size, flags | M_BESTFIT, &addr))
415 rv = kmem_back_domain(domain, kernel_object, addr, size, flags);
417 vmem_free(arena, addr, size);
420 return (addr);
434 vm_offset_t addr;
439 addr = kmem_malloc_domain(domain, size, flags);
440 if (addr != 0)
444 return (addr);
454 kmem_back_domain(int domain, vm_object_t object, vm_offset_t addr,
465 offset = addr - VM_MIN_KERNEL_ADDRESS;
489 kmem_unback(object, addr, i);
500 pmap_enter(kernel_pmap, addr + i, m, prot,
516 kmem_back(vm_object_t object, vm_offset_t addr, vm_size_t size, int flags)
524 for (start = addr, end = addr + size; addr < end; addr = next) {
530 domain = (addr >> KVA_QUANTUM_SHIFT) % vm_ndomains;
533 next = roundup2(addr + 1, KVA_QUANTUM);
540 rv = kmem_back_domain(domain, object, addr, next - addr, flags);
542 kmem_unback(object, start, addr - start);
559 _kmem_unback(vm_object_t object, vm_offset_t addr, vm_size_t size)
571 pmap_remove(kernel_pmap, addr, addr + size);
572 offset = addr - VM_MIN_KERNEL_ADDRESS;
593 kmem_unback(vm_object_t object, vm_offset_t addr, vm_size_t size)
596 (void)_kmem_unback(object, addr, size);
606 kmem_free(vm_offset_t addr, vm_size_t size)
611 arena = _kmem_unback(kernel_object, addr, size);
613 vmem_free(arena, addr, size);
627 vm_offset_t addr;
639 addr = vm_map_findspace(map, vm_map_min(map), size);
640 if (addr + size <= vm_map_max(map))
651 vm_map_insert(map, NULL, 0, addr, addr + size, VM_PROT_RW, VM_PROT_RW,
654 return (addr);
664 kmap_free_wakeup(vm_map_t map, vm_offset_t addr, vm_size_t size)
668 (void) vm_map_delete(map, trunc_page(addr), round_page(addr + size));
679 vm_offset_t addr, i;
687 addr = kva_alloc(ZERO_REGION_SIZE);
693 pmap_qenter(addr + i, &m, 1);
694 pmap_protect(kernel_pmap, addr, addr + ZERO_REGION_SIZE, VM_PROT_READ);
696 zero_region = (const void *)addr;
705 vm_offset_t addr;
711 addr = vm_map_min(kernel_map);
712 result = vm_map_find(kernel_map, NULL, 0, &addr, size, 0,
717 *addrp = addr;