/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_evict.c | 57 obj = i915_gem_object_create_internal(ggtt->vm.i915, 75 count, ggtt->vm.total / PAGE_SIZE); 77 if (list_empty(&ggtt->vm.bound_list)) { 89 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link) 104 i915_gem_drain_freed_objects(ggtt->vm.i915); 121 mutex_lock(&ggtt->vm.mutex); 122 err = i915_gem_evict_something(&ggtt->vm, NULL, 126 mutex_unlock(&ggtt->vm.mutex); 136 mutex_lock(&ggtt->vm.mutex); 137 err = i915_gem_evict_something(&ggtt->vm, NUL [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm_pt.c | 150 * @vm: amdgpu_vm structure 157 struct amdgpu_vm *vm, uint64_t start, 162 cursor->entry = &vm->root; 279 * @vm: amdgpu_vm structure 286 struct amdgpu_vm *vm, 293 amdgpu_vm_pt_start(adev, vm, 0, cursor); 340 #define for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry) \ 341 for (amdgpu_vm_pt_first_dfs((adev), (vm), (start), &(cursor)), \ 350 * @vm: VM to clear BO from 359 int amdgpu_vm_pt_clear(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 156 amdgpu_vm_pt_start(struct amdgpu_device *adev, struct amdgpu_vm *vm, uint64_t start, struct amdgpu_vm_pt_cursor *cursor) argument 285 amdgpu_vm_pt_first_dfs(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start, struct amdgpu_vm_pt_cursor *cursor) argument 446 amdgpu_vm_pt_create(struct amdgpu_device *adev, struct amdgpu_vm *vm, int level, bool immediate, struct amdgpu_bo_vm **vmbo, int32_t xcp_id) argument 539 amdgpu_vm_pt_alloc(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *cursor, bool immediate) argument 606 struct amdgpu_vm *vm; local 634 amdgpu_vm_pt_free_dfs(struct amdgpu_device *adev, struct amdgpu_vm *vm, struct amdgpu_vm_pt_cursor *start, bool unlocked) argument 668 amdgpu_vm_pt_free_root(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument 686 struct amdgpu_vm *vm = params->vm; local 933 struct amdgpu_vm *vm = params->vm; local 999 amdgpu_vm_pt_map_tables(struct amdgpu_device *adev, struct amdgpu_vm *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | access_tracking_perf_test.c | 95 static uint64_t lookup_pfn(int pagemap_fd, struct kvm_vm *vm, uint64_t gva) argument 97 uint64_t hva = (uint64_t) addr_gva2hva(vm, gva); 126 static void mark_vcpu_memory_idle(struct kvm_vm *vm, argument 150 uint64_t pfn = lookup_pfn(pagemap_fd, vm, gva); 229 struct kvm_vm *vm = memstress_args.vm; local 240 mark_vcpu_memory_idle(vm, vcpu_args); 262 static void run_iteration(struct kvm_vm *vm, int nr_vcpus, const char *description) argument 282 static void access_memory(struct kvm_vm *vm, int nr_vcpus, argument 285 memstress_set_write_percent(vm, (acces 290 mark_memory_idle(struct kvm_vm *vm, int nr_vcpus) argument 306 struct kvm_vm *vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | dirty_log_page_splitting_test.c | 41 static void get_page_stats(struct kvm_vm *vm, struct kvm_page_stats *stats, const char *stage) argument 43 stats->pages_4k = vm_get_stat(vm, "pages_4k"); 44 stats->pages_2m = vm_get_stat(vm, "pages_2m"); 45 stats->pages_1g = vm_get_stat(vm, "pages_1g"); 53 static void run_vcpu_iteration(struct kvm_vm *vm) argument 89 struct kvm_vm *vm; local 102 vm = memstress_create_vm(mode, VCPUS, guest_percpu_mem_size, 105 guest_num_pages = (VCPUS * guest_percpu_mem_size) >> vm->page_shift; 116 vm_enable_cap(vm, KVM_CAP_MANUAL_DIRTY_LOG_PROTECT2, 128 run_vcpu_iteration(vm); [all...] |
H A D | svm_int_ctl_test.c | 89 struct kvm_vm *vm; local 94 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 96 vm_init_descriptor_tables(vm); 99 vm_install_exception_handler(vm, VINTR_IRQ_NUMBER, vintr_irq_handler); 100 vm_install_exception_handler(vm, INTR_IRQ_NUMBER, intr_irq_handler); 102 vcpu_alloc_svm(vm, &svm_gva); 119 kvm_vm_free(vm);
|
H A D | vmx_apic_access_test.c | 81 struct kvm_vm *vm; local 85 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 87 high_gpa = (vm->max_gfn - 1) << vm->page_shift; 89 vmx = vcpu_alloc_vmx(vm, &vmx_pages_gva); 90 prepare_virtualize_apic_accesses(vmx, vm); 122 kvm_vm_free(vm);
|
H A D | platform_info_test.c | 41 vm_enable_cap(vcpu->vm, KVM_CAP_MSR_PLATFORM_INFO, true); 56 vm_enable_cap(vcpu->vm, KVM_CAP_MSR_PLATFORM_INFO, false); 64 struct kvm_vm *vm; local 69 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 78 kvm_vm_free(vm);
|
H A D | ucna_injection_test.c | 220 struct kvm_vm *vm = vcpu->vm; local 222 params->p_i_ucna_rcvd = (uint64_t *)addr_gva2hva(vm, (uint64_t)&i_ucna_rcvd); 223 params->p_i_ucna_addr = (uint64_t *)addr_gva2hva(vm, (uint64_t)&i_ucna_addr); 224 params->p_ucna_addr = (uint64_t *)addr_gva2hva(vm, (uint64_t)&ucna_addr); 225 params->p_ucna_addr2 = (uint64_t *)addr_gva2hva(vm, (uint64_t)&ucna_addr2); 256 static struct kvm_vcpu *create_vcpu_with_mce_cap(struct kvm_vm *vm, uint32_t vcpuid, argument 259 struct kvm_vcpu *vcpu = vm_vcpu_add(vm, vcpuid, guest_code); 267 struct kvm_vm *vm; local 274 vm [all...] |
H A D | cr4_cpuid_sync_test.c | 53 struct kvm_vm *vm; local 59 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 83 kvm_vm_free(vm);
|
H A D | vmx_close_while_nested_test.c | 52 struct kvm_vm *vm; local 56 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 59 vcpu_alloc_vmx(vm, &vmx_pages_gva);
|
H A D | hwcr_msr_test.c | 37 struct kvm_vm *vm; local 41 vm = vm_create_with_one_vcpu(&vcpu, NULL); 46 kvm_vm_free(vm);
|
H A D | svm_nested_soft_inject_test.c | 146 struct kvm_vm *vm; local 153 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 155 vm_init_descriptor_tables(vm); 158 vm_install_exception_handler(vm, NMI_VECTOR, guest_nmi_handler); 159 vm_install_exception_handler(vm, BP_VECTOR, guest_bp_handler); 160 vm_install_exception_handler(vm, INT_NR, guest_int_handler); 162 vcpu_alloc_svm(vm, &svm_gva); 167 idt_alt_vm = vm_vaddr_alloc_page(vm); 168 idt_alt = addr_gva2hva(vm, idt_alt_vm); 169 idt = addr_gva2hva(vm, v [all...] |
H A D | kvm_clock_test.c | 77 static void setup_clock(struct kvm_vm *vm, struct test_case *test_case) argument 102 vm_ioctl(vm, KVM_SET_CLOCK, &data); 108 struct kvm_vm *vm = vcpu->vm; local 113 setup_clock(vm, &test_cases[i]); 115 vm_ioctl(vm, KVM_GET_CLOCK, &start); 118 vm_ioctl(vm, KVM_GET_CLOCK, &end); 140 struct kvm_vm *vm; local 148 vm = vm_create_with_one_vcpu(&vcpu, guest_main); 150 pvti_gva = vm_vaddr_alloc(vm, getpagesiz [all...] |
H A D | xss_msr_test.c | 20 struct kvm_vm *vm; local 26 vm = vm_create_with_one_vcpu(&vcpu, NULL); 55 kvm_vm_free(vm);
|
/linux-master/tools/testing/selftests/kvm/s390x/ |
H A D | debug_test.c | 24 struct kvm_vm *vm; local 27 vm = vm_create_with_one_vcpu(vcpu, guest_code); 28 lowcore = addr_gpa2hva(vm, 0); 39 return vm; 46 struct kvm_vm *vm; local 48 vm = test_step_int_1(&vcpu, guest_code, new_psw_off, new_psw); 52 kvm_vm_free(vm); 83 struct kvm_vm *vm; local 85 vm = test_step_int_1(&vcpu, test_step_pgm_diag_guest_code, 95 kvm_vm_free(vm); [all...] |
/linux-master/drivers/gpu/drm/kmb/ |
H A D | kmb_crtc.c | 73 struct kmb_crtc_timing vm; local 90 vm.vfront_porch = 2; 91 vm.vback_porch = 2; 92 vm.vsync_len = 8; 93 vm.hfront_porch = 0; 94 vm.hback_porch = 0; 95 vm.hsync_len = 28; 99 m->crtc_vdisplay, vm.vback_porch, vm.vfront_porch, 100 vm [all...] |
/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | arch_timer.c | 170 static void test_init_timer_irq(struct kvm_vm *vm) argument 178 sync_global_to_guest(vm, ptimer_irq); 179 sync_global_to_guest(vm, vtimer_irq); 188 struct kvm_vm *vm; local 192 vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus); 194 vm_init_descriptor_tables(vm); 195 vm_install_exception_handler(vm, VECTOR_IRQ_CURRENT, guest_irq_handler); 203 vm_ioctl(vm, KVM_ARM_SET_COUNTER_OFFSET, &offset); 211 test_init_timer_irq(vm); 212 gic_fd = vgic_v3_setup(vm, nr_vcpu 221 test_vm_cleanup(struct kvm_vm *vm) argument [all...] |
H A D | psci_test.c | 79 struct kvm_vm *vm; local 81 vm = vm_create(2); 83 vm_ioctl(vm, KVM_ARM_PREFERRED_TARGET, &init); 86 *source = aarch64_vcpu_add(vm, 0, &init, guest_code); 87 *target = aarch64_vcpu_add(vm, 1, &init, guest_code); 89 return vm; 136 struct kvm_vm *vm; local 139 vm = setup_vm(guest_test_cpu_on, &source, &target); 154 kvm_vm_free(vm); 173 struct kvm_vm *vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | memstress.c | 62 void memstress_setup_ept(struct vmx_pages *vmx, struct kvm_vm *vm) argument 66 prepare_eptp(vmx, vm, 0); 73 nested_identity_map_1g(vmx, vm, 0, 0x100000000ULL); 77 nested_identity_map_1g(vmx, vm, start, end - start); 80 void memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu *vcpus[]) argument 91 vmx = vcpu_alloc_vmx(vm, &vmx_gva); 94 memstress_setup_ept(vmx, vm);
|
H A D | svm.c | 23 * vm - The VM to allocate guest-virtual addresses in. 32 vcpu_alloc_svm(struct kvm_vm *vm, vm_vaddr_t *p_svm_gva) argument 34 vm_vaddr_t svm_gva = vm_vaddr_alloc_page(vm); 35 struct svm_test_data *svm = addr_gva2hva(vm, svm_gva); 37 svm->vmcb = (void *)vm_vaddr_alloc_page(vm); 38 svm->vmcb_hva = addr_gva2hva(vm, (uintptr_t)svm->vmcb); 39 svm->vmcb_gpa = addr_gva2gpa(vm, (uintptr_t)svm->vmcb); 41 svm->save_area = (void *)vm_vaddr_alloc_page(vm); 42 svm->save_area_hva = addr_gva2hva(vm, (uintptr_t)svm->save_area); 43 svm->save_area_gpa = addr_gva2gpa(vm, (uintptr_ [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_evict.c | 46 static int ggtt_flush(struct i915_address_space *vm) argument 48 struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); 125 * @vm: address space to evict from 148 i915_gem_evict_something(struct i915_address_space *vm, argument 164 lockdep_assert_held(&vm->mutex); 165 trace_i915_gem_evict(vm, min_size, alignment, flags); 183 drm_mm_scan_init_with_range(&scan, &vm->mm, 187 if (i915_is_ggtt(vm)) { 188 struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); 193 intel_gt_retire_requests(vm 328 i915_gem_evict_for_node(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_mm_node *target, unsigned int flags) argument 458 i915_gem_evict_vm(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_i915_gem_object **busy_bo) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_exec_queue.h | 17 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, 22 struct xe_vm *vm, 63 void xe_exec_queue_last_fence_put(struct xe_exec_queue *e, struct xe_vm *vm); 66 struct xe_vm *vm); 67 void xe_exec_queue_last_fence_set(struct xe_exec_queue *e, struct xe_vm *vm,
|
/linux-master/tools/testing/selftests/net/ |
H A D | test_vxlan_vnifiltering.sh | 14 # | vm-11 netns | | vm-21 netns | 76 # | vm-12 netns | |vm-22 netns | 188 # example: setup-vm <hvid> <vmid> <brid> <VATTRS> <mcast_for_bum> 194 # setup-vm 1 11 1 \ 197 setup-vm() { 210 # create vm namespace and interfaces and connect to hypervisor 216 ip link set $vmvethif netns ${vm[$vmid]} 218 ip -netns ${vm[ [all...] |
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_mic.c | 103 struct videomode vm; member in struct:exynos_mic 157 struct videomode vm = mic->vm; local 160 reg = MIC_V_PULSE_WIDTH(vm.vsync_len) + 161 MIC_V_PERIOD_LINE(vm.vsync_len + vm.vactive + 162 vm.vback_porch + vm.vfront_porch); 165 reg = MIC_VBP_SIZE(vm.vback_porch) + 166 MIC_VFP_SIZE(vm 181 struct videomode *vm = &mic->vm; local 192 struct videomode vm = mic->vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/include/aarch64/ |
H A D | vgic.h | 19 int vgic_v3_setup(struct kvm_vm *vm, unsigned int nr_vcpus, uint32_t nr_irqs, 27 void kvm_arm_irq_line(struct kvm_vm *vm, uint32_t intid, int level); 28 int _kvm_arm_irq_line(struct kvm_vm *vm, uint32_t intid, int level);
|