/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
H A D | mock_gtt.c | 32 static void mock_insert_page(struct i915_address_space *vm, argument 40 static void mock_insert_entries(struct i915_address_space *vm, argument 59 static void mock_cleanup(struct i915_address_space *vm) argument 63 static void mock_clear_range(struct i915_address_space *vm, argument 76 ppgtt->vm.gt = &i915->gt; 77 ppgtt->vm.i915 = i915; 78 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); 79 ppgtt->vm.file = ERR_PTR(-ENODEV); 81 i915_address_space_init(&ppgtt->vm, VM_CLASS_PPGTT); 83 ppgtt->vm [all...] |
H A D | i915_gem_gtt.c | 164 if (!ppgtt->vm.allocate_va_range) 175 limit = min(ppgtt->vm.total, limit); 179 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, size); 191 ppgtt->vm.clear_range(&ppgtt->vm, 0, size); 196 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 211 i915_vm_put(&ppgtt->vm); 215 static int lowlevel_hole(struct i915_address_space *vm, argument 327 close_object_list(struct list_head *objects, struct i915_address_space *vm) argument 348 fill_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 559 walk_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 642 pot_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 717 drunk_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 826 __shrink_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 901 shrink_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 923 shrink_boom(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument 1001 exercise_ppgtt(struct drm_i915_private *dev_priv, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument 1081 exercise_ggtt(struct drm_i915_private *i915, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument 1250 exercise_mock(struct drm_i915_private *i915, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument 1256 struct i915_address_space *vm; local 1786 struct i915_address_space *vm; local [all...] |
H A D | i915_gem_evict.c | 60 obj = i915_gem_object_create_internal(ggtt->vm.i915, 78 count, ggtt->vm.total / PAGE_SIZE); 104 if (list_empty(&ggtt->vm.bound_list)) { 116 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link) 131 i915_gem_drain_freed_objects(ggtt->vm.i915); 148 mutex_lock(&ggtt->vm.mutex); 149 err = i915_gem_evict_something(&ggtt->vm, 153 mutex_unlock(&ggtt->vm.mutex); 163 mutex_lock(&ggtt->vm.mutex); 164 err = i915_gem_evict_something(&ggtt->vm, [all...] |
/netbsd-current/sys/kern/ |
H A D | subr_vmem.c | 98 #define VMEM_CONDVAR_INIT(vm, wchan) cv_init(&vm->vm_cv, wchan) 99 #define VMEM_CONDVAR_DESTROY(vm) cv_destroy(&vm->vm_cv) 100 #define VMEM_CONDVAR_WAIT(vm) cv_wait(&vm->vm_cv, &vm->vm_lock) 101 #define VMEM_CONDVAR_BROADCAST(vm) cv_broadcast(&vm->vm_cv) 108 #define VMEM_CONDVAR_INIT(vm, wcha 243 bt_refill_locked(vmem_t *vm) argument 297 bt_refill(vmem_t *vm) argument 308 bt_alloc(vmem_t *vm, vm_flag_t flags) argument 341 bt_free(vmem_t *vm, bt_t *bt) argument 351 bt_freetrim(vmem_t *vm, int freelimit) argument 394 vmem_add_bts(vmem_t *vm, struct vmem_btag *bts, unsigned int nbts) argument 419 bt_freehead_tofree(vmem_t *vm, vmem_size_t size) argument 443 bt_freehead_toalloc(vmem_t *vm, vmem_size_t size, vm_flag_t strat) argument 465 bt_hashhead(vmem_t *vm, vmem_addr_t addr) argument 477 bt_lookupbusy(vmem_t *vm, vmem_addr_t addr) argument 493 bt_rembusy(vmem_t *vm, bt_t *bt) argument 503 bt_insbusy(vmem_t *vm, bt_t *bt) argument 520 bt_remseg(vmem_t *vm, bt_t *bt) argument 527 bt_insseg(vmem_t *vm, bt_t *bt, bt_t *prev) argument 534 bt_insseg_tail(vmem_t *vm, bt_t *bt) argument 541 bt_remfree(vmem_t *vm, bt_t *bt) argument 550 bt_insfree(vmem_t *vm, bt_t *bt) argument 603 vmem_t *vm = qc->qc_vmem; local 616 vmem_t *vm = qc->qc_vmem; local 622 qc_init(vmem_t *vm, size_t qcache_max, int ipl) argument 674 qc_destroy(vmem_t *vm) argument 713 vmem_subsystem_init(vmem_t *vm) argument 733 vmem_add1(vmem_t *vm, vmem_addr_t addr, vmem_size_t size, vm_flag_t flags, int spanbttype) argument 772 vmem_destroy1(vmem_t *vm) argument 803 vmem_import(vmem_t *vm, vmem_size_t size, vm_flag_t flags) argument 843 vmem_rehash(vmem_t *vm, size_t newhashsize, vm_flag_t flags) argument 964 vmem_init(vmem_t *vm, const char *name, vmem_addr_t base, vmem_size_t size, vmem_size_t quantum, vmem_import_t *importfn, vmem_release_t *releasefn, vmem_t *arg, vmem_size_t qcache_max, vm_flag_t flags, int ipl) argument 1086 vmem_destroy(vmem_t *vm) argument 1099 vmem_roundup_size(vmem_t *vm, vmem_size_t size) argument 1110 vmem_alloc(vmem_t *vm, vmem_size_t size, vm_flag_t flags, vmem_addr_t *addrp) argument 1152 vmem_xalloc_addr(vmem_t *vm, const vmem_addr_t addr, const vmem_size_t size, vm_flag_t flags) argument 1172 vmem_xalloc(vmem_t *vm, const vmem_size_t size0, vmem_size_t align, const vmem_size_t phase, const vmem_size_t nocross, const vmem_addr_t minaddr, const vmem_addr_t maxaddr, const vm_flag_t flags, vmem_addr_t *addrp) argument 1370 vmem_free(vmem_t *vm, vmem_addr_t addr, vmem_size_t size) argument 1392 vmem_xfree(vmem_t *vm, vmem_addr_t addr, vmem_size_t size) argument 1415 vmem_xfreeall(vmem_t *vm) argument 1441 vmem_xfree_bt(vmem_t *vm, bt_t *bt) argument 1506 vmem_add(vmem_t *vm, vmem_addr_t addr, vmem_size_t size, vm_flag_t flags) argument 1523 vmem_size(vmem_t *vm, int typemask) argument 1549 vmem_t *vm; local 1660 vmem_whatis_lookup(vmem_t *vm, uintptr_t addr) argument 1679 vmem_t *vm; local 1698 const vmem_t *vm; local 1708 const vmem_t *vm = (const void *)addr; local 1733 vmem_check_sanity(vmem_t *vm) argument 1768 vmem_check(vmem_t *vm) argument 1783 vmem_t *vm; local [all...] |
H A D | sys_process.c | 154 struct vmspace *vm; local 169 vm = p->p_vmspace; 171 if ((l->l_flag & LW_WEXIT) || vm->vm_refcnt < 1) 177 error = uvm_io(&vm->vm_map, uio, pax_mprotect_prot(l)); 183 uvmspace_free(vm);
|
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | intel_gtt.c | 57 static struct page *vm_alloc_page(struct i915_address_space *vm, gfp_t gfp) argument 62 if (I915_SELFTEST_ONLY(should_fail(&vm->fault_attr, 1))) 63 i915_gem_shrink_all(vm->i915); 65 page = stash_pop_page(&vm->free_pages); 69 if (!vm->pt_kmap_wc) 73 page = stash_pop_page(&vm->i915->mm.wc_stash); 101 stash_push_pagevec(&vm->i915->mm.wc_stash, &stack); 105 stash_push_pagevec(&vm->free_pages, &stack); 117 static void vm_free_pages_release(struct i915_address_space *vm, argument 120 struct pagevec *pvec = &vm 160 vm_free_page(struct i915_address_space *vm, struct page *page) argument 179 __i915_vm_close(struct i915_address_space *vm) argument 201 i915_address_space_fini(struct i915_address_space *vm) argument 218 struct i915_address_space *vm = local 229 struct i915_address_space *vm = local 238 i915_address_space_init(struct i915_address_space *vm, int subclass) argument 279 __setup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p, gfp_t gfp) argument 341 setup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p) argument 346 cleanup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p) argument 364 setup_scratch_page(struct i915_address_space *vm, gfp_t gfp) argument 477 cleanup_scratch_page(struct i915_address_space *vm) argument 494 free_scratch(struct i915_address_space *vm) argument [all...] |
H A D | gen8_ppgtt.c | 35 struct drm_i915_private *i915 = ppgtt->vm.i915; 36 struct intel_uncore *uncore = ppgtt->vm.gt->uncore; 47 if (i915_vm_is_4lvl(&ppgtt->vm)) { 126 gen8_pd_top_count(const struct i915_address_space *vm) argument 128 unsigned int shift = __gen8_pte_shift(vm->top); 129 return (vm->total + (1ull << shift) - 1) >> shift; 133 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) argument 135 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); 137 if (vm->top == 2) 140 return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm 144 gen8_pdp_for_page_address(struct i915_address_space * const vm, const u64 addr) argument 149 __gen8_ppgtt_cleanup(struct i915_address_space *vm, struct i915_page_directory *pd, int count, int lvl) argument 168 gen8_ppgtt_cleanup(struct i915_address_space *vm) argument 179 __gen8_ppgtt_clear(struct i915_address_space * const vm, struct i915_page_directory * const pd, u64 start, const u64 end, int lvl) argument 241 gen8_ppgtt_clear(struct i915_address_space *vm, u64 start, u64 length) argument 256 __gen8_ppgtt_alloc(struct i915_address_space * const vm, struct i915_page_directory * const pd, u64 * const start, const u64 end, int lvl) argument 359 gen8_ppgtt_alloc(struct i915_address_space *vm, u64 start, u64 length) argument 633 gen8_ppgtt_insert(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level cache_level, u32 flags) argument 658 gen8_init_scratch(struct i915_address_space *vm) argument 705 struct i915_address_space *vm = &ppgtt->vm; local 729 gen8_alloc_top_pd(struct i915_address_space *vm) argument [all...] |
H A D | intel_ggtt.c | 53 struct drm_i915_private *i915 = ggtt->vm.i915; 55 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); 57 ggtt->vm.is_ggtt = true; 60 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); 63 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; 69 ggtt->vm.cleanup(&ggtt->vm); 81 ggtt->vm.cleanup(&ggtt->vm); 135 struct drm_i915_private *i915 = ggtt->vm 217 gen8_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 unused) argument 239 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument 290 gen6_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 flags) argument 319 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument 367 nop_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 372 gen8_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 401 bxt_vtd_ggtt_wa(struct i915_address_space *vm) argument 414 struct i915_address_space *vm; member in struct:insert_page 430 bxt_vtd_ggtt_insert_page__BKL(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 unused) argument 442 struct i915_address_space *vm; member in struct:insert_entries 458 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument 469 struct i915_address_space *vm; member in struct:clear_range 484 bxt_vtd_ggtt_clear_range__BKL(struct i915_address_space *vm, u64 start, u64 length) argument 493 gen6_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 526 i915_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level cache_level, u32 unused) argument 538 i915_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level cache_level, u32 unused) argument 550 i915_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 730 struct i915_address_space *vm = vma->vm; local 736 struct i915_address_space *vm = local 988 gen6_gmch_remove(struct i915_address_space *vm) argument 1226 i915_gmch_remove(struct i915_address_space *vm) argument [all...] |
H A D | intel_gtt.h | 77 #define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT) 279 * Since the vm may be shared between multiple contexts, we count how 317 int (*allocate_va_range)(struct i915_address_space *vm, 319 void (*clear_range)(struct i915_address_space *vm, 321 void (*insert_page)(struct i915_address_space *vm, 326 void (*insert_entries)(struct i915_address_space *vm, 330 void (*cleanup)(struct i915_address_space *vm); 347 struct i915_address_space vm; member in struct:i915_ggtt 409 struct i915_address_space vm; member in struct:i915_ppgtt 414 #define i915_is_ggtt(vm) ((v 417 i915_vm_is_4lvl(const struct i915_address_space *vm) argument 423 i915_vm_has_scratch_64K(struct i915_address_space *vm) argument 429 i915_vm_has_cache_coloring(struct i915_address_space *vm) argument 435 i915_vm_to_ggtt(struct i915_address_space *vm) argument 443 i915_vm_to_ppgtt(struct i915_address_space *vm) argument 451 i915_vm_get(struct i915_address_space *vm) argument 459 i915_vm_put(struct i915_address_space *vm) argument 465 i915_vm_open(struct i915_address_space *vm) argument 473 i915_vm_tryopen(struct i915_address_space *vm) argument 484 i915_vm_close(struct i915_address_space *vm) argument [all...] |
H A D | intel_ppgtt.c | 18 struct i915_page_table *alloc_pt(struct i915_address_space *vm) argument 26 if (unlikely(setup_page_dma(vm, &pt->base))) { 47 struct i915_page_directory *alloc_pd(struct i915_address_space *vm) argument 55 if (unlikely(setup_page_dma(vm, px_base(pd)))) { 64 void free_pd(struct i915_address_space *vm, struct i915_page_dma *pd) argument 66 cleanup_page_dma(vm, pd); 163 trace_i915_ppgtt_create(&ppgtt->vm); 176 err = vma->vm->allocate_va_range(vma->vm, 190 vma->vm [all...] |
H A D | gen6_ppgtt.c | 89 static void gen6_ppgtt_clear_range(struct i915_address_space *vm, argument 92 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); 94 const gen6_pte_t scratch_pte = vm->scratch[0].encode; 105 GEM_BUG_ON(px_base(pt) == px_base(&vm->scratch[1])); 128 static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, argument 133 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); 138 const u32 pte_encode = vm->pte_encode(0, cache_level, flags); 142 GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]); 210 gen6_ggtt_invalidate(ppgtt->base.vm.gt->ggtt); 216 static int gen6_alloc_va_range(struct i915_address_space *vm, argument 275 struct i915_address_space * const vm = &ppgtt->base.vm; local 311 gen6_ppgtt_cleanup(struct i915_address_space *vm) argument [all...] |
/netbsd-current/sys/external/bsd/drm2/include/linux/sched/ |
H A D | mm.h | 39 mmgrab(struct vmspace *vm) argument 41 uvmspace_addref(vm); 45 mmdrop(struct vmspace *vm) argument 47 uvmspace_free(vm);
|
/netbsd-current/external/gpl3/gdb/dist/sim/ppc/ |
H A D | hw_vm.c | 31 vm - virtual memory device for user simulation modes 82 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 85 vm->stack_base = device_find_integer_property(me, "stack-base"); 86 vm->stack_bound = (vm->stack_base 88 vm->stack_lower_limit = vm->stack_bound; 89 vm->heap_base = 0; 90 vm->heap_bound = 0; 91 vm 113 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 137 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 224 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 265 hw_vm_device *vm = ZALLOC(hw_vm_device); local [all...] |
H A D | vm.h | 24 typedef struct _vm vm; typedef in typeref:struct:_vm 32 (vm *) vm_create 37 (vm *memory); 41 (vm *memory); 130 /* update vm data structures due to a synchronization point */ 134 (vm *memory, 143 /* update vm data structures due to a TLB operation */ 147 (vm *memory, 152 (vm *memory);
|
/netbsd-current/external/gpl3/gdb.old/dist/sim/ppc/ |
H A D | hw_vm.c | 31 vm - virtual memory device for user simulation modes 82 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 85 vm->stack_base = device_find_integer_property(me, "stack-base"); 86 vm->stack_bound = (vm->stack_base 88 vm->stack_lower_limit = vm->stack_bound; 89 vm->heap_base = 0; 90 vm->heap_bound = 0; 91 vm 113 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 137 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 224 hw_vm_device *vm = (hw_vm_device*)device_data(me); local 265 hw_vm_device *vm = ZALLOC(hw_vm_device); local [all...] |
H A D | vm.h | 24 typedef struct _vm vm; typedef in typeref:struct:_vm 32 (vm *) vm_create 37 (vm *memory); 41 (vm *memory); 130 /* update vm data structures due to a synchronization point */ 134 (vm *memory, 143 /* update vm data structures due to a TLB operation */ 147 (vm *memory, 152 (vm *memory);
|
/netbsd-current/external/lgpl3/gmp/dist/mini-gmp/tests/ |
H A D | t-lucm.c | 34 mpz_t m, vr, qr, vm, qm, vt; local 42 mpz_init (vm); 54 resm = mpz_lucas_mod (vm, qm, Q, b0, m); 58 if (resm != 0 || mpz_cmp_ui (vm, 0) != 0) 63 dump ("vm", vm); 72 mpz_abs (vm, vm); 77 (mpz_cmp (vm, vr) != 0 && mpz_cmp (vm, v [all...] |
/netbsd-current/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_vm.c | 85 * radeon_vm_manager_init - init the vm manager 89 * Init the vm manager (cayman+). 107 * radeon_vm_manager_fini - tear down the vm manager 127 * radeon_vm_get_bos - add the vm BOs to a validation list 129 * @vm: vm providing the BOs 136 struct radeon_vm *vm, 142 list = kvmalloc_array(vm->max_pde_used + 2, 147 /* add the vm page table to the list */ 148 list[0].robj = vm 135 radeon_vm_get_bos(struct radeon_device *rdev, struct radeon_vm *vm, struct list_head *head) argument 184 radeon_vm_grab_id(struct radeon_device *rdev, struct radeon_vm *vm, int ring) argument 243 radeon_vm_flush(struct radeon_device *rdev, struct radeon_vm *vm, int ring, struct radeon_fence *updates) argument 275 radeon_vm_fence(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_fence *fence) argument 300 radeon_vm_bo_find(struct radeon_vm *vm, struct radeon_bo *bo) argument 326 radeon_vm_bo_add(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_bo *bo) argument 459 struct radeon_vm *vm = bo_va->vm; local 648 radeon_vm_update_page_directory(struct radeon_device *rdev, struct radeon_vm *vm) argument 821 radeon_vm_update_ptes(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_ib *ib, uint64_t start, uint64_t end, uint64_t dst, uint32_t flags) argument 893 radeon_vm_fence_pts(struct radeon_vm *vm, uint64_t start, uint64_t end, struct radeon_fence *fence) argument 923 struct radeon_vm *vm = bo_va->vm; local 1055 radeon_vm_clear_freed(struct radeon_device *rdev, struct radeon_vm *vm) argument 1093 radeon_vm_clear_invalids(struct radeon_device *rdev, struct radeon_vm *vm) argument 1129 struct radeon_vm *vm = bo_va->vm; local 1182 radeon_vm_init(struct radeon_device *rdev, struct radeon_vm *vm) argument 1242 radeon_vm_fini(struct radeon_device *rdev, struct radeon_vm *vm) argument [all...] |
/netbsd-current/usr.sbin/videomode/ |
H A D | videomode.c | 114 struct grfvideo_mode vm; local 126 vm.mode_num = 0; 127 if (ioctl(grffd, GRFGETVMODE, &vm) == 0) 128 dump_vm(&vm); 136 vm.mode_num = m; 137 if (ioctl(grffd, GRFGETVMODE, &vm) == -1) 139 dump_vm(&vm); 155 dump_vm(struct grfvideo_mode *vm) argument 157 (void)printf("%d: %s\n", vm->mode_num, vm [all...] |
/netbsd-current/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 92 * vm eviction_lock can be taken in MMU notifiers. Make sure no reclaim-FS 96 static inline void amdgpu_vm_eviction_lock(struct amdgpu_vm *vm) argument 98 mutex_lock(&vm->eviction_lock); 99 vm->saved_flags = memalloc_nofs_save(); 102 static inline int amdgpu_vm_eviction_trylock(struct amdgpu_vm *vm) argument 104 if (mutex_trylock(&vm->eviction_lock)) { 105 vm->saved_flags = memalloc_nofs_save(); 111 static inline void amdgpu_vm_eviction_unlock(struct amdgpu_vm *vm) argument 113 memalloc_nofs_restore(vm->saved_flags); 114 mutex_unlock(&vm 235 struct amdgpu_vm *vm = vm_bo->vm; local 324 amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base, struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument 397 amdgpu_vm_pt_start(struct amdgpu_device *adev, struct amdgpu_vm *vm, uint64_t start, struct amdgpu_vm_pt_cursor *cursor) argument 523 amdgpu_vm_pt_first_dfs(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start, struct amdgpu_vm_pt_cursor *cursor) argument 591 amdgpu_vm_get_pd_bo(struct amdgpu_vm *vm, struct list_head *validated, struct amdgpu_bo_list_entry *entry) argument 626 struct amdgpu_vm *vm = bo_base->vm; local 642 amdgpu_vm_move_to_lru_tail(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 686 amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, int (*validate)(void *p, struct amdgpu_bo *bo), void *param) argument 730 amdgpu_vm_ready(struct amdgpu_vm *vm) argument 748 amdgpu_vm_clear_bo(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo *bo, bool direct) argument 868 amdgpu_vm_bo_param(struct amdgpu_device *adev, struct amdgpu_vm *vm, int level, bool direct, struct amdgpu_bo_param *bp) argument 904 amdgpu_vm_alloc_pts(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *cursor, bool direct) argument 978 amdgpu_vm_free_pts(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start) argument 1195 amdgpu_vm_bo_find(struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument 1254 amdgpu_vm_update_pde(struct amdgpu_vm_update_params *params, struct amdgpu_vm *vm, struct amdgpu_vm_pt *entry) argument 1280 amdgpu_vm_invalidate_pds(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 1303 amdgpu_vm_update_pdes(struct amdgpu_device *adev, struct amdgpu_vm *vm, bool direct) argument 1581 amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev, struct amdgpu_vm *vm, bool direct, struct dma_fence *exclusive, uint64_t start, uint64_t last, uint64_t flags, uint64_t addr, bus_dma_segment_t *pages_addr, struct dma_fence **fence) argument 1647 amdgpu_vm_bo_split_mapping(struct amdgpu_device *adev, struct dma_fence *exclusive, bus_dma_segment_t *pages_addr, struct amdgpu_vm *vm, struct amdgpu_bo_va_mapping *mapping, uint64_t flags, struct amdgpu_device *bo_adev, struct drm_mm_node *nodes, struct dma_fence **fence) argument 1778 struct amdgpu_vm *vm = bo_va->base.vm; local 1961 amdgpu_vm_free_mapping(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo_va_mapping *mapping, struct dma_fence *fence) argument 1979 amdgpu_vm_prt_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 2024 amdgpu_vm_clear_freed(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct dma_fence **fence) argument 2076 amdgpu_vm_handle_moved(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 2133 amdgpu_vm_bo_add(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument 2176 struct amdgpu_vm *vm = bo_va->base.vm; local 2217 struct amdgpu_vm *vm = bo_va->base.vm; local 2339 struct amdgpu_vm *vm = bo_va->base.vm; local 2388 amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev, struct amdgpu_vm *vm, uint64_t saddr, uint64_t size) argument 2491 amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm *vm, uint64_t addr) argument 2505 amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket) argument 2542 struct amdgpu_vm *vm = bo_va->base.vm; local 2643 struct amdgpu_vm *vm = bo_base->vm; local 2784 amdgpu_vm_wait_idle(struct amdgpu_vm *vm, long timeout) argument 2811 amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm, int vm_context, unsigned int pasid) argument 2955 amdgpu_vm_check_clean_reserved(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 2994 amdgpu_vm_make_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm, unsigned int pasid) argument 3094 amdgpu_vm_release_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 3116 amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 3287 struct amdgpu_vm *vm; local 3304 amdgpu_vm_set_task_info(struct amdgpu_vm *vm) argument 3343 struct amdgpu_vm *vm; local [all...] |
/netbsd-current/tests/bin/sh/ |
H A D | t_patterns.sh | 726 vm() function 736 vm abc \# a bc; vm aaab \# a aab; vm aaab \## 'a*a' b # 3 737 vm aaab % ab aa; vm xawab %% 'a*ab' x; vm abcd \# xyz abcd 738 vm file.c % .c 'f le' IFS=i ; vm file.c % .c file IFS=i Q 739 vm fil [all...] |
/netbsd-current/sys/uvm/ |
H A D | uvm_unix.c | 44 * uvm_unix.c: traditional sbrk/grow interface to vm. 74 struct vmspace *vm = p->p_vmspace; local 79 obreak = (vaddr_t)vm->vm_daddr; 88 obreak = round_page(obreak + ptoa(vm->vm_dsize)); 105 error = uvm_map(&vm->vm_map, &obreak, nbreak - obreak, NULL, 119 vm->vm_dsize += atop(nbreak - obreak); 121 uvm_deallocate(&vm->vm_map, nbreak, obreak - nbreak); 122 vm->vm_dsize -= atop(obreak - nbreak); 136 struct vmspace *vm = p->p_vmspace; local 143 if (sp < (vaddr_t)vm [all...] |
/netbsd-current/lib/libkvm/ |
H A D | kvm_m68k.c | 88 struct vmstate *vm; local 90 vm = (struct vmstate *)_kvm_malloc(kd, sizeof (*vm)); 91 if (vm == 0) 94 kd->vmst = vm; 107 vm->ops = nop->ops; 112 for (vm->pgshift = 0; (1 << vm->pgshift) < h->page_size; vm->pgshift++) 114 if ((1 << vm [all...] |
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/ |
H A D | i915_gem_evict.c | 75 * @vm: address space to evict from 97 i915_gem_evict_something(struct i915_address_space *vm, argument 111 lockdep_assert_held(&vm->mutex); 112 trace_i915_gem_evict(vm, min_size, alignment, flags); 130 drm_mm_scan_init_with_range(&scan, &vm->mm, 134 intel_gt_retire_requests(vm->gt); 139 list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) { 167 list_move_tail(&vma->vm_link, &vm->bound_list); 188 if (!i915_is_ggtt(vm) || flags & PIN_NONBLOCK) 207 ret = ggtt_flush(vm 257 i915_gem_evict_for_node(struct i915_address_space *vm, struct drm_mm_node *target, unsigned int flags) argument 367 i915_gem_evict_vm(struct i915_address_space *vm) argument [all...] |
/netbsd-current/external/gpl3/gcc/dist/libstdc++-v3/src/c++17/ryu/ |
H A D | f2s.c | 66 uint32_t vr, vp, vm; local 78 vm = mulPow5InvDivPow2(mm, q, i); 81 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm); 83 if (q != 0 && (vp - 1) / 10 <= vm / 10) { 109 vm = mulPow5divPow2(mm, (uint32_t) i, j); 113 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm); 115 if (q != 0 && (vp - 1) / 10 <= vm / 10) { 120 // {vr,vp,vm} is trailing zeros if {mv,mp,mm} has at least q trailing 0 bits. 139 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm); 140 printf("vm i [all...] |