Searched refs:vm (Results 1 - 25 of 315) sorted by relevance

1234567891011>>

/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/selftests/
H A Dmock_gtt.c32 static void mock_insert_page(struct i915_address_space *vm, argument
40 static void mock_insert_entries(struct i915_address_space *vm, argument
59 static void mock_cleanup(struct i915_address_space *vm) argument
63 static void mock_clear_range(struct i915_address_space *vm, argument
76 ppgtt->vm.gt = &i915->gt;
77 ppgtt->vm.i915 = i915;
78 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE);
79 ppgtt->vm.file = ERR_PTR(-ENODEV);
81 i915_address_space_init(&ppgtt->vm, VM_CLASS_PPGTT);
83 ppgtt->vm
[all...]
H A Di915_gem_gtt.c164 if (!ppgtt->vm.allocate_va_range)
175 limit = min(ppgtt->vm.total, limit);
179 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, size);
191 ppgtt->vm.clear_range(&ppgtt->vm, 0, size);
196 err = ppgtt->vm.allocate_va_range(&ppgtt->vm,
211 i915_vm_put(&ppgtt->vm);
215 static int lowlevel_hole(struct i915_address_space *vm, argument
327 close_object_list(struct list_head *objects, struct i915_address_space *vm) argument
348 fill_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
559 walk_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
642 pot_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
717 drunk_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
826 __shrink_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
901 shrink_hole(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
923 shrink_boom(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time) argument
1001 exercise_ppgtt(struct drm_i915_private *dev_priv, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument
1081 exercise_ggtt(struct drm_i915_private *i915, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument
1250 exercise_mock(struct drm_i915_private *i915, int (*func)(struct i915_address_space *vm, u64 hole_start, u64 hole_end, unsigned long end_time)) argument
1256 struct i915_address_space *vm; local
1786 struct i915_address_space *vm; local
[all...]
H A Di915_gem_evict.c60 obj = i915_gem_object_create_internal(ggtt->vm.i915,
78 count, ggtt->vm.total / PAGE_SIZE);
104 if (list_empty(&ggtt->vm.bound_list)) {
116 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link)
131 i915_gem_drain_freed_objects(ggtt->vm.i915);
148 mutex_lock(&ggtt->vm.mutex);
149 err = i915_gem_evict_something(&ggtt->vm,
153 mutex_unlock(&ggtt->vm.mutex);
163 mutex_lock(&ggtt->vm.mutex);
164 err = i915_gem_evict_something(&ggtt->vm,
[all...]
/netbsd-current/sys/kern/
H A Dsubr_vmem.c98 #define VMEM_CONDVAR_INIT(vm, wchan) cv_init(&vm->vm_cv, wchan)
99 #define VMEM_CONDVAR_DESTROY(vm) cv_destroy(&vm->vm_cv)
100 #define VMEM_CONDVAR_WAIT(vm) cv_wait(&vm->vm_cv, &vm->vm_lock)
101 #define VMEM_CONDVAR_BROADCAST(vm) cv_broadcast(&vm->vm_cv)
108 #define VMEM_CONDVAR_INIT(vm, wcha
243 bt_refill_locked(vmem_t *vm) argument
297 bt_refill(vmem_t *vm) argument
308 bt_alloc(vmem_t *vm, vm_flag_t flags) argument
341 bt_free(vmem_t *vm, bt_t *bt) argument
351 bt_freetrim(vmem_t *vm, int freelimit) argument
394 vmem_add_bts(vmem_t *vm, struct vmem_btag *bts, unsigned int nbts) argument
419 bt_freehead_tofree(vmem_t *vm, vmem_size_t size) argument
443 bt_freehead_toalloc(vmem_t *vm, vmem_size_t size, vm_flag_t strat) argument
465 bt_hashhead(vmem_t *vm, vmem_addr_t addr) argument
477 bt_lookupbusy(vmem_t *vm, vmem_addr_t addr) argument
493 bt_rembusy(vmem_t *vm, bt_t *bt) argument
503 bt_insbusy(vmem_t *vm, bt_t *bt) argument
520 bt_remseg(vmem_t *vm, bt_t *bt) argument
527 bt_insseg(vmem_t *vm, bt_t *bt, bt_t *prev) argument
534 bt_insseg_tail(vmem_t *vm, bt_t *bt) argument
541 bt_remfree(vmem_t *vm, bt_t *bt) argument
550 bt_insfree(vmem_t *vm, bt_t *bt) argument
603 vmem_t *vm = qc->qc_vmem; local
616 vmem_t *vm = qc->qc_vmem; local
622 qc_init(vmem_t *vm, size_t qcache_max, int ipl) argument
674 qc_destroy(vmem_t *vm) argument
713 vmem_subsystem_init(vmem_t *vm) argument
733 vmem_add1(vmem_t *vm, vmem_addr_t addr, vmem_size_t size, vm_flag_t flags, int spanbttype) argument
772 vmem_destroy1(vmem_t *vm) argument
803 vmem_import(vmem_t *vm, vmem_size_t size, vm_flag_t flags) argument
843 vmem_rehash(vmem_t *vm, size_t newhashsize, vm_flag_t flags) argument
964 vmem_init(vmem_t *vm, const char *name, vmem_addr_t base, vmem_size_t size, vmem_size_t quantum, vmem_import_t *importfn, vmem_release_t *releasefn, vmem_t *arg, vmem_size_t qcache_max, vm_flag_t flags, int ipl) argument
1086 vmem_destroy(vmem_t *vm) argument
1099 vmem_roundup_size(vmem_t *vm, vmem_size_t size) argument
1110 vmem_alloc(vmem_t *vm, vmem_size_t size, vm_flag_t flags, vmem_addr_t *addrp) argument
1152 vmem_xalloc_addr(vmem_t *vm, const vmem_addr_t addr, const vmem_size_t size, vm_flag_t flags) argument
1172 vmem_xalloc(vmem_t *vm, const vmem_size_t size0, vmem_size_t align, const vmem_size_t phase, const vmem_size_t nocross, const vmem_addr_t minaddr, const vmem_addr_t maxaddr, const vm_flag_t flags, vmem_addr_t *addrp) argument
1370 vmem_free(vmem_t *vm, vmem_addr_t addr, vmem_size_t size) argument
1392 vmem_xfree(vmem_t *vm, vmem_addr_t addr, vmem_size_t size) argument
1415 vmem_xfreeall(vmem_t *vm) argument
1441 vmem_xfree_bt(vmem_t *vm, bt_t *bt) argument
1506 vmem_add(vmem_t *vm, vmem_addr_t addr, vmem_size_t size, vm_flag_t flags) argument
1523 vmem_size(vmem_t *vm, int typemask) argument
1549 vmem_t *vm; local
1660 vmem_whatis_lookup(vmem_t *vm, uintptr_t addr) argument
1679 vmem_t *vm; local
1698 const vmem_t *vm; local
1708 const vmem_t *vm = (const void *)addr; local
1733 vmem_check_sanity(vmem_t *vm) argument
1768 vmem_check(vmem_t *vm) argument
1783 vmem_t *vm; local
[all...]
H A Dsys_process.c154 struct vmspace *vm; local
169 vm = p->p_vmspace;
171 if ((l->l_flag & LW_WEXIT) || vm->vm_refcnt < 1)
177 error = uvm_io(&vm->vm_map, uio, pax_mprotect_prot(l));
183 uvmspace_free(vm);
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gt/
H A Dintel_gtt.c57 static struct page *vm_alloc_page(struct i915_address_space *vm, gfp_t gfp) argument
62 if (I915_SELFTEST_ONLY(should_fail(&vm->fault_attr, 1)))
63 i915_gem_shrink_all(vm->i915);
65 page = stash_pop_page(&vm->free_pages);
69 if (!vm->pt_kmap_wc)
73 page = stash_pop_page(&vm->i915->mm.wc_stash);
101 stash_push_pagevec(&vm->i915->mm.wc_stash, &stack);
105 stash_push_pagevec(&vm->free_pages, &stack);
117 static void vm_free_pages_release(struct i915_address_space *vm, argument
120 struct pagevec *pvec = &vm
160 vm_free_page(struct i915_address_space *vm, struct page *page) argument
179 __i915_vm_close(struct i915_address_space *vm) argument
201 i915_address_space_fini(struct i915_address_space *vm) argument
218 struct i915_address_space *vm = local
229 struct i915_address_space *vm = local
238 i915_address_space_init(struct i915_address_space *vm, int subclass) argument
279 __setup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p, gfp_t gfp) argument
341 setup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p) argument
346 cleanup_page_dma(struct i915_address_space *vm, struct i915_page_dma *p) argument
364 setup_scratch_page(struct i915_address_space *vm, gfp_t gfp) argument
477 cleanup_scratch_page(struct i915_address_space *vm) argument
494 free_scratch(struct i915_address_space *vm) argument
[all...]
H A Dgen8_ppgtt.c35 struct drm_i915_private *i915 = ppgtt->vm.i915;
36 struct intel_uncore *uncore = ppgtt->vm.gt->uncore;
47 if (i915_vm_is_4lvl(&ppgtt->vm)) {
126 gen8_pd_top_count(const struct i915_address_space *vm) argument
128 unsigned int shift = __gen8_pte_shift(vm->top);
129 return (vm->total + (1ull << shift) - 1) >> shift;
133 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) argument
135 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm);
137 if (vm->top == 2)
140 return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm
144 gen8_pdp_for_page_address(struct i915_address_space * const vm, const u64 addr) argument
149 __gen8_ppgtt_cleanup(struct i915_address_space *vm, struct i915_page_directory *pd, int count, int lvl) argument
168 gen8_ppgtt_cleanup(struct i915_address_space *vm) argument
179 __gen8_ppgtt_clear(struct i915_address_space * const vm, struct i915_page_directory * const pd, u64 start, const u64 end, int lvl) argument
241 gen8_ppgtt_clear(struct i915_address_space *vm, u64 start, u64 length) argument
256 __gen8_ppgtt_alloc(struct i915_address_space * const vm, struct i915_page_directory * const pd, u64 * const start, const u64 end, int lvl) argument
359 gen8_ppgtt_alloc(struct i915_address_space *vm, u64 start, u64 length) argument
633 gen8_ppgtt_insert(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level cache_level, u32 flags) argument
658 gen8_init_scratch(struct i915_address_space *vm) argument
705 struct i915_address_space *vm = &ppgtt->vm; local
729 gen8_alloc_top_pd(struct i915_address_space *vm) argument
[all...]
H A Dintel_ggtt.c53 struct drm_i915_private *i915 = ggtt->vm.i915;
55 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT);
57 ggtt->vm.is_ggtt = true;
60 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915);
63 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust;
69 ggtt->vm.cleanup(&ggtt->vm);
81 ggtt->vm.cleanup(&ggtt->vm);
135 struct drm_i915_private *i915 = ggtt->vm
217 gen8_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 unused) argument
239 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument
290 gen6_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 flags) argument
319 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument
367 nop_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument
372 gen8_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument
401 bxt_vtd_ggtt_wa(struct i915_address_space *vm) argument
414 struct i915_address_space *vm; member in struct:insert_page
430 bxt_vtd_ggtt_insert_page__BKL(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level level, u32 unused) argument
442 struct i915_address_space *vm; member in struct:insert_entries
458 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level level, u32 flags) argument
469 struct i915_address_space *vm; member in struct:clear_range
484 bxt_vtd_ggtt_clear_range__BKL(struct i915_address_space *vm, u64 start, u64 length) argument
493 gen6_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument
526 i915_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, enum i915_cache_level cache_level, u32 unused) argument
538 i915_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma *vma, enum i915_cache_level cache_level, u32 unused) argument
550 i915_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument
730 struct i915_address_space *vm = vma->vm; local
736 struct i915_address_space *vm = local
988 gen6_gmch_remove(struct i915_address_space *vm) argument
1226 i915_gmch_remove(struct i915_address_space *vm) argument
[all...]
H A Dintel_gtt.h77 #define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT)
279 * Since the vm may be shared between multiple contexts, we count how
317 int (*allocate_va_range)(struct i915_address_space *vm,
319 void (*clear_range)(struct i915_address_space *vm,
321 void (*insert_page)(struct i915_address_space *vm,
326 void (*insert_entries)(struct i915_address_space *vm,
330 void (*cleanup)(struct i915_address_space *vm);
347 struct i915_address_space vm; member in struct:i915_ggtt
409 struct i915_address_space vm; member in struct:i915_ppgtt
414 #define i915_is_ggtt(vm) ((v
417 i915_vm_is_4lvl(const struct i915_address_space *vm) argument
423 i915_vm_has_scratch_64K(struct i915_address_space *vm) argument
429 i915_vm_has_cache_coloring(struct i915_address_space *vm) argument
435 i915_vm_to_ggtt(struct i915_address_space *vm) argument
443 i915_vm_to_ppgtt(struct i915_address_space *vm) argument
451 i915_vm_get(struct i915_address_space *vm) argument
459 i915_vm_put(struct i915_address_space *vm) argument
465 i915_vm_open(struct i915_address_space *vm) argument
473 i915_vm_tryopen(struct i915_address_space *vm) argument
484 i915_vm_close(struct i915_address_space *vm) argument
[all...]
H A Dintel_ppgtt.c18 struct i915_page_table *alloc_pt(struct i915_address_space *vm) argument
26 if (unlikely(setup_page_dma(vm, &pt->base))) {
47 struct i915_page_directory *alloc_pd(struct i915_address_space *vm) argument
55 if (unlikely(setup_page_dma(vm, px_base(pd)))) {
64 void free_pd(struct i915_address_space *vm, struct i915_page_dma *pd) argument
66 cleanup_page_dma(vm, pd);
163 trace_i915_ppgtt_create(&ppgtt->vm);
176 err = vma->vm->allocate_va_range(vma->vm,
190 vma->vm
[all...]
H A Dgen6_ppgtt.c89 static void gen6_ppgtt_clear_range(struct i915_address_space *vm, argument
92 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm));
94 const gen6_pte_t scratch_pte = vm->scratch[0].encode;
105 GEM_BUG_ON(px_base(pt) == px_base(&vm->scratch[1]));
128 static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, argument
133 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm);
138 const u32 pte_encode = vm->pte_encode(0, cache_level, flags);
142 GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]);
210 gen6_ggtt_invalidate(ppgtt->base.vm.gt->ggtt);
216 static int gen6_alloc_va_range(struct i915_address_space *vm, argument
275 struct i915_address_space * const vm = &ppgtt->base.vm; local
311 gen6_ppgtt_cleanup(struct i915_address_space *vm) argument
[all...]
/netbsd-current/sys/external/bsd/drm2/include/linux/sched/
H A Dmm.h39 mmgrab(struct vmspace *vm) argument
41 uvmspace_addref(vm);
45 mmdrop(struct vmspace *vm) argument
47 uvmspace_free(vm);
/netbsd-current/external/gpl3/gdb/dist/sim/ppc/
H A Dhw_vm.c31 vm - virtual memory device for user simulation modes
82 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
85 vm->stack_base = device_find_integer_property(me, "stack-base");
86 vm->stack_bound = (vm->stack_base
88 vm->stack_lower_limit = vm->stack_bound;
89 vm->heap_base = 0;
90 vm->heap_bound = 0;
91 vm
113 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
137 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
224 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
265 hw_vm_device *vm = ZALLOC(hw_vm_device); local
[all...]
H A Dvm.h24 typedef struct _vm vm; typedef in typeref:struct:_vm
32 (vm *) vm_create
37 (vm *memory);
41 (vm *memory);
130 /* update vm data structures due to a synchronization point */
134 (vm *memory,
143 /* update vm data structures due to a TLB operation */
147 (vm *memory,
152 (vm *memory);
/netbsd-current/external/gpl3/gdb.old/dist/sim/ppc/
H A Dhw_vm.c31 vm - virtual memory device for user simulation modes
82 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
85 vm->stack_base = device_find_integer_property(me, "stack-base");
86 vm->stack_bound = (vm->stack_base
88 vm->stack_lower_limit = vm->stack_bound;
89 vm->heap_base = 0;
90 vm->heap_bound = 0;
91 vm
113 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
137 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
224 hw_vm_device *vm = (hw_vm_device*)device_data(me); local
265 hw_vm_device *vm = ZALLOC(hw_vm_device); local
[all...]
H A Dvm.h24 typedef struct _vm vm; typedef in typeref:struct:_vm
32 (vm *) vm_create
37 (vm *memory);
41 (vm *memory);
130 /* update vm data structures due to a synchronization point */
134 (vm *memory,
143 /* update vm data structures due to a TLB operation */
147 (vm *memory,
152 (vm *memory);
/netbsd-current/external/lgpl3/gmp/dist/mini-gmp/tests/
H A Dt-lucm.c34 mpz_t m, vr, qr, vm, qm, vt; local
42 mpz_init (vm);
54 resm = mpz_lucas_mod (vm, qm, Q, b0, m);
58 if (resm != 0 || mpz_cmp_ui (vm, 0) != 0)
63 dump ("vm", vm);
72 mpz_abs (vm, vm);
77 (mpz_cmp (vm, vr) != 0 && mpz_cmp (vm, v
[all...]
/netbsd-current/sys/external/bsd/drm2/dist/drm/radeon/
H A Dradeon_vm.c85 * radeon_vm_manager_init - init the vm manager
89 * Init the vm manager (cayman+).
107 * radeon_vm_manager_fini - tear down the vm manager
127 * radeon_vm_get_bos - add the vm BOs to a validation list
129 * @vm: vm providing the BOs
136 struct radeon_vm *vm,
142 list = kvmalloc_array(vm->max_pde_used + 2,
147 /* add the vm page table to the list */
148 list[0].robj = vm
135 radeon_vm_get_bos(struct radeon_device *rdev, struct radeon_vm *vm, struct list_head *head) argument
184 radeon_vm_grab_id(struct radeon_device *rdev, struct radeon_vm *vm, int ring) argument
243 radeon_vm_flush(struct radeon_device *rdev, struct radeon_vm *vm, int ring, struct radeon_fence *updates) argument
275 radeon_vm_fence(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_fence *fence) argument
300 radeon_vm_bo_find(struct radeon_vm *vm, struct radeon_bo *bo) argument
326 radeon_vm_bo_add(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_bo *bo) argument
459 struct radeon_vm *vm = bo_va->vm; local
648 radeon_vm_update_page_directory(struct radeon_device *rdev, struct radeon_vm *vm) argument
821 radeon_vm_update_ptes(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_ib *ib, uint64_t start, uint64_t end, uint64_t dst, uint32_t flags) argument
893 radeon_vm_fence_pts(struct radeon_vm *vm, uint64_t start, uint64_t end, struct radeon_fence *fence) argument
923 struct radeon_vm *vm = bo_va->vm; local
1055 radeon_vm_clear_freed(struct radeon_device *rdev, struct radeon_vm *vm) argument
1093 radeon_vm_clear_invalids(struct radeon_device *rdev, struct radeon_vm *vm) argument
1129 struct radeon_vm *vm = bo_va->vm; local
1182 radeon_vm_init(struct radeon_device *rdev, struct radeon_vm *vm) argument
1242 radeon_vm_fini(struct radeon_device *rdev, struct radeon_vm *vm) argument
[all...]
/netbsd-current/usr.sbin/videomode/
H A Dvideomode.c114 struct grfvideo_mode vm; local
126 vm.mode_num = 0;
127 if (ioctl(grffd, GRFGETVMODE, &vm) == 0)
128 dump_vm(&vm);
136 vm.mode_num = m;
137 if (ioctl(grffd, GRFGETVMODE, &vm) == -1)
139 dump_vm(&vm);
155 dump_vm(struct grfvideo_mode *vm) argument
157 (void)printf("%d: %s\n", vm->mode_num, vm
[all...]
/netbsd-current/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
H A Damdgpu_vm.c92 * vm eviction_lock can be taken in MMU notifiers. Make sure no reclaim-FS
96 static inline void amdgpu_vm_eviction_lock(struct amdgpu_vm *vm) argument
98 mutex_lock(&vm->eviction_lock);
99 vm->saved_flags = memalloc_nofs_save();
102 static inline int amdgpu_vm_eviction_trylock(struct amdgpu_vm *vm) argument
104 if (mutex_trylock(&vm->eviction_lock)) {
105 vm->saved_flags = memalloc_nofs_save();
111 static inline void amdgpu_vm_eviction_unlock(struct amdgpu_vm *vm) argument
113 memalloc_nofs_restore(vm->saved_flags);
114 mutex_unlock(&vm
235 struct amdgpu_vm *vm = vm_bo->vm; local
324 amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base, struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument
397 amdgpu_vm_pt_start(struct amdgpu_device *adev, struct amdgpu_vm *vm, uint64_t start, struct amdgpu_vm_pt_cursor *cursor) argument
523 amdgpu_vm_pt_first_dfs(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start, struct amdgpu_vm_pt_cursor *cursor) argument
591 amdgpu_vm_get_pd_bo(struct amdgpu_vm *vm, struct list_head *validated, struct amdgpu_bo_list_entry *entry) argument
626 struct amdgpu_vm *vm = bo_base->vm; local
642 amdgpu_vm_move_to_lru_tail(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
686 amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, int (*validate)(void *p, struct amdgpu_bo *bo), void *param) argument
730 amdgpu_vm_ready(struct amdgpu_vm *vm) argument
748 amdgpu_vm_clear_bo(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo *bo, bool direct) argument
868 amdgpu_vm_bo_param(struct amdgpu_device *adev, struct amdgpu_vm *vm, int level, bool direct, struct amdgpu_bo_param *bp) argument
904 amdgpu_vm_alloc_pts(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *cursor, bool direct) argument
978 amdgpu_vm_free_pts(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start) argument
1195 amdgpu_vm_bo_find(struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument
1254 amdgpu_vm_update_pde(struct amdgpu_vm_update_params *params, struct amdgpu_vm *vm, struct amdgpu_vm_pt *entry) argument
1280 amdgpu_vm_invalidate_pds(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
1303 amdgpu_vm_update_pdes(struct amdgpu_device *adev, struct amdgpu_vm *vm, bool direct) argument
1581 amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev, struct amdgpu_vm *vm, bool direct, struct dma_fence *exclusive, uint64_t start, uint64_t last, uint64_t flags, uint64_t addr, bus_dma_segment_t *pages_addr, struct dma_fence **fence) argument
1647 amdgpu_vm_bo_split_mapping(struct amdgpu_device *adev, struct dma_fence *exclusive, bus_dma_segment_t *pages_addr, struct amdgpu_vm *vm, struct amdgpu_bo_va_mapping *mapping, uint64_t flags, struct amdgpu_device *bo_adev, struct drm_mm_node *nodes, struct dma_fence **fence) argument
1778 struct amdgpu_vm *vm = bo_va->base.vm; local
1961 amdgpu_vm_free_mapping(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo_va_mapping *mapping, struct dma_fence *fence) argument
1979 amdgpu_vm_prt_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
2024 amdgpu_vm_clear_freed(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct dma_fence **fence) argument
2076 amdgpu_vm_handle_moved(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
2133 amdgpu_vm_bo_add(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_bo *bo) argument
2176 struct amdgpu_vm *vm = bo_va->base.vm; local
2217 struct amdgpu_vm *vm = bo_va->base.vm; local
2339 struct amdgpu_vm *vm = bo_va->base.vm; local
2388 amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev, struct amdgpu_vm *vm, uint64_t saddr, uint64_t size) argument
2491 amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm *vm, uint64_t addr) argument
2505 amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket) argument
2542 struct amdgpu_vm *vm = bo_va->base.vm; local
2643 struct amdgpu_vm *vm = bo_base->vm; local
2784 amdgpu_vm_wait_idle(struct amdgpu_vm *vm, long timeout) argument
2811 amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm, int vm_context, unsigned int pasid) argument
2955 amdgpu_vm_check_clean_reserved(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
2994 amdgpu_vm_make_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm, unsigned int pasid) argument
3094 amdgpu_vm_release_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
3116 amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument
3287 struct amdgpu_vm *vm; local
3304 amdgpu_vm_set_task_info(struct amdgpu_vm *vm) argument
3343 struct amdgpu_vm *vm; local
[all...]
/netbsd-current/tests/bin/sh/
H A Dt_patterns.sh726 vm() function
736 vm abc \# a bc; vm aaab \# a aab; vm aaab \## 'a*a' b # 3
737 vm aaab % ab aa; vm xawab %% 'a*ab' x; vm abcd \# xyz abcd
738 vm file.c % .c 'f le' IFS=i ; vm file.c % .c file IFS=i Q
739 vm fil
[all...]
/netbsd-current/sys/uvm/
H A Duvm_unix.c44 * uvm_unix.c: traditional sbrk/grow interface to vm.
74 struct vmspace *vm = p->p_vmspace; local
79 obreak = (vaddr_t)vm->vm_daddr;
88 obreak = round_page(obreak + ptoa(vm->vm_dsize));
105 error = uvm_map(&vm->vm_map, &obreak, nbreak - obreak, NULL,
119 vm->vm_dsize += atop(nbreak - obreak);
121 uvm_deallocate(&vm->vm_map, nbreak, obreak - nbreak);
122 vm->vm_dsize -= atop(obreak - nbreak);
136 struct vmspace *vm = p->p_vmspace; local
143 if (sp < (vaddr_t)vm
[all...]
/netbsd-current/lib/libkvm/
H A Dkvm_m68k.c88 struct vmstate *vm; local
90 vm = (struct vmstate *)_kvm_malloc(kd, sizeof (*vm));
91 if (vm == 0)
94 kd->vmst = vm;
107 vm->ops = nop->ops;
112 for (vm->pgshift = 0; (1 << vm->pgshift) < h->page_size; vm->pgshift++)
114 if ((1 << vm
[all...]
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/
H A Di915_gem_evict.c75 * @vm: address space to evict from
97 i915_gem_evict_something(struct i915_address_space *vm, argument
111 lockdep_assert_held(&vm->mutex);
112 trace_i915_gem_evict(vm, min_size, alignment, flags);
130 drm_mm_scan_init_with_range(&scan, &vm->mm,
134 intel_gt_retire_requests(vm->gt);
139 list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) {
167 list_move_tail(&vma->vm_link, &vm->bound_list);
188 if (!i915_is_ggtt(vm) || flags & PIN_NONBLOCK)
207 ret = ggtt_flush(vm
257 i915_gem_evict_for_node(struct i915_address_space *vm, struct drm_mm_node *target, unsigned int flags) argument
367 i915_gem_evict_vm(struct i915_address_space *vm) argument
[all...]
/netbsd-current/external/gpl3/gcc/dist/libstdc++-v3/src/c++17/ryu/
H A Df2s.c66 uint32_t vr, vp, vm; local
78 vm = mulPow5InvDivPow2(mm, q, i);
81 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm);
83 if (q != 0 && (vp - 1) / 10 <= vm / 10) {
109 vm = mulPow5divPow2(mm, (uint32_t) i, j);
113 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm);
115 if (q != 0 && (vp - 1) / 10 <= vm / 10) {
120 // {vr,vp,vm} is trailing zeros if {mv,mp,mm} has at least q trailing 0 bits.
139 printf("V+=%u\nV =%u\nV-=%u\n", vp, vr, vm);
140 printf("vm i
[all...]

Completed in 271 milliseconds

1234567891011>>