/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | hypercalls.c | 160 vm_userspace_mem_region_add(vcpu->vm, VM_MEM_SRC_ANONYMOUS, ST_GPA_BASE, 1, gpages, 0); 236 struct kvm_vm *vm; local 238 vm = vm_create_with_one_vcpu(vcpu, guest_code); 242 return vm; 245 static void test_guest_stage(struct kvm_vm **vm, struct kvm_vcpu **vcpu) argument 253 sync_global_to_guest(*vm, stage); 261 kvm_vm_free(*vm); 262 *vm = test_vm_create(vcpu); 275 struct kvm_vm *vm; local 279 vm [all...] |
H A D | vgic_irq.c | 83 static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, 516 static void kvm_irq_line_check(struct kvm_vm *vm, uint32_t intid, int level, argument 522 kvm_arm_irq_line(vm, intid, level); 528 ret = _kvm_arm_irq_line(vm, intid, level); 559 static void kvm_set_gsi_routing_irqchip_check(struct kvm_vm *vm, argument 574 kvm_gsi_routing_write(vm, routing); 576 ret = _kvm_gsi_routing_write(vm, routing); 604 static void kvm_routing_and_irqfd_check(struct kvm_vm *vm, argument 621 kvm_set_gsi_routing_irqchip_check(vm, intid, num, 641 vm_ioctl(vm, KVM_IRQF 670 struct kvm_vm *vm = vcpu->vm; local 720 kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, struct kvm_inject_args *args) argument 743 struct kvm_vm *vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | userspace_msr_exit_test.c | 519 struct kvm_vm *vm = vcpu->vm; local 523 sync_global_to_guest(vm, fep_available); 527 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MSR, KVM_MSR_EXIT_REASON_FILTER); 532 vm_ioctl(vm, KVM_X86_SET_MSR_FILTER, &filter_allow); 534 vm_init_descriptor_tables(vm); 537 vm_install_exception_handler(vm, GP_VECTOR, guest_gp_handler); 556 vm_install_exception_handler(vm, GP_VECTOR, guest_fep_gp_handler); 587 vm_ioctl(vcpu->vm, KVM_X86_SET_MSR_FILTER, &no_filter_deny); 637 struct kvm_vm *vm local 678 struct kvm_vm *vm = vcpu->vm; local 709 run_user_space_msr_flag_test(struct kvm_vm *vm) argument 726 run_msr_filter_flag_test(struct kvm_vm *vm) argument 766 struct kvm_vm *vm = vcpu->vm; local [all...] |
H A D | private_mem_conversions_test.c | 298 struct kvm_vm *vm = vcpu->vm; local 305 vm_guest_mem_fallocate(vm, gpa, size, map_shared); 308 vm_set_memory_attributes(vm, gpa, size, 319 struct kvm_vm *vm = vcpu->vm; local 349 for (i = 0; i < size; i += vm->page_size) { 350 size_t nr_bytes = min_t(size_t, vm->page_size, size - i); 351 uint8_t *hva = addr_gpa2hva(vm, gpa + i); 383 struct kvm_vm *vm; local [all...] |
H A D | xapic_state_test.c | 166 struct kvm_vm *vm; local 169 vm = vm_create_with_vcpus(NR_VCPUS, NULL, vcpus); 170 vm_enable_cap(vm, KVM_CAP_X2APIC_API, KVM_X2APIC_API_USE_32BIT_IDS); 185 kvm_vm_free(vm); 194 struct kvm_vm *vm; local 196 vm = vm_create_with_one_vcpu(&x.vcpu, x2apic_guest_code); 198 kvm_vm_free(vm); 205 vm = vm_create_with_one_vcpu(&x.vcpu, xapic_guest_code); 210 virt_pg_map(vm, APIC_DEFAULT_GPA, APIC_DEFAULT_GPA); 212 kvm_vm_free(vm); [all...] |
H A D | xen_shinfo_test.c | 411 struct kvm_vm *vm = (struct kvm_vm *)arg; local 436 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_activate_gfn); 438 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_deactivate_gfn); 441 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_activate_hva); 443 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_deactivate_hva); 454 struct kvm_vm *vm; local 473 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 476 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 478 virt_map(vm, SHINFO_REGION_GVA, SHINFO_REGION_GPA, 3); 480 shinfo = addr_gpa2hva(vm, SHINFO_VADD [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | steal_time.c | 74 sync_global_to_guest(vcpu->vm, st_gva[i]); 83 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 85 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); 171 struct kvm_vm *vm = vcpu->vm; local 185 sync_global_to_guest(vm, st_gva[i]); 198 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 200 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); 284 st_gpa[i] = addr_gva2gpa(vcpu->vm, (vm_vaddr_t)st_gva[i]); 285 sync_global_to_guest(vcpu->vm, st_gv 289 steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 344 struct kvm_vm *vm; local [all...] |
H A D | guest_memfd_test.c | 125 static void test_create_guest_memfd_invalid(struct kvm_vm *vm) argument 133 fd = __vm_create_guest_memfd(vm, size, 0); 140 fd = __vm_create_guest_memfd(vm, page_size, flag); 147 static void test_create_guest_memfd_multiple(struct kvm_vm *vm) argument 152 fd1 = __vm_create_guest_memfd(vm, 4096, 0); 159 fd2 = __vm_create_guest_memfd(vm, 8192, 0); 180 struct kvm_vm *vm; local 187 vm = vm_create_barebones(); 189 test_create_guest_memfd_invalid(vm); 190 test_create_guest_memfd_multiple(vm); [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_csa.c | 65 int amdgpu_map_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 74 r = amdgpu_vm_lock_pd(vm, &exec, 0); 84 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); 105 int amdgpu_unmap_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 114 r = amdgpu_vm_lock_pd(vm, &exec, 0);
|
H A D | amdgpu_seq64.c | 55 * @vm: vm pointer 63 int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 77 r = amdgpu_vm_lock_pd(vm, &exec, 0); 85 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); 122 struct amdgpu_vm *vm; local 134 vm = &fpriv->vm; 138 r = amdgpu_vm_lock_pd(vm, &exec, 0);
|
H A D | amdgpu_ids.c | 262 * @vm: vm to allocate id for 270 static int amdgpu_vmid_grab_reserved(struct amdgpu_vm *vm, argument 280 bool needs_flush = vm->use_cpu_for_update; 281 uint64_t updates = amdgpu_vm_tlb_seq(vm); 285 if ((*id)->owner != vm->immediate.fence_context || 323 * @vm: vm to allocate id for 331 static int amdgpu_vmid_grab_used(struct amdgpu_vm *vm, argument 341 uint64_t updates = amdgpu_vm_tlb_seq(vm); 394 amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct dma_fence **fence) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_exec_queue.c | 37 struct xe_vm *vm, 89 if (vm) 90 q->vm = xe_vm_get(vm); 102 if (q->vm) 103 xe_vm_put(q->vm); 113 err = xe_lrc_init(q->lrc + i, q->hwe, q, q->vm, SZ_16K); 123 * Normally the user vm holds an rpm ref to keep the device 124 * awake, and the context holds a ref for the vm, however for 125 * some engines we use the kernels migrate vm underneat 36 __xe_exec_queue_alloc(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument 141 xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument 173 xe_exec_queue_create_class(struct xe_device *xe, struct xe_gt *gt, struct xe_vm *vm, enum xe_engine_class class, u32 flags) argument 555 struct xe_vm *vm, *migrate_vm; local 820 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue *q, struct xe_vm *vm) argument 834 xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) argument 867 xe_exec_queue_last_fence_get(struct xe_exec_queue *q, struct xe_vm *vm) argument 892 xe_exec_queue_last_fence_set(struct xe_exec_queue *q, struct xe_vm *vm, struct dma_fence *fence) argument [all...] |
H A D | xe_trace.h | 86 __field(struct xe_vm *, vm) 92 __entry->vm = bo->vm; 95 TP_printk("size=%zu, flags=0x%02x, vm=%p", 96 __entry->size, __entry->flags, __entry->vm) 487 TP_PROTO(struct xe_vm *vm), 488 TP_ARGS(vm), 491 __field(struct xe_vm *, vm) 496 __entry->vm = vm; [all...] |
H A D | xe_bo.h | 17 * xe_vm_assert_held(vm) - Assert that the vm's reservation object is held. 18 * @vm: The vm 20 #define xe_vm_assert_held(vm) dma_resv_assert_held(xe_vm_resv(vm)) 105 struct xe_tile *tile, struct xe_vm *vm, 109 struct xe_vm *vm, size_t size, 112 struct xe_vm *vm, size_t size, 115 struct xe_vm *vm, size_ [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_trace.h | 64 __field(struct i915_address_space *, vm) 72 __entry->vm = vma->vm; 78 TP_printk("obj=%p, offset=0x%016llx size=0x%llx%s vm=%p", 81 __entry->vm) 90 __field(struct i915_address_space *, vm) 97 __entry->vm = vma->vm; 102 TP_printk("obj=%p, offset=0x%016llx size=0x%llx vm=%p", 103 __entry->obj, __entry->offset, __entry->size, __entry->vm) [all...] |
H A D | i915_vgpu.c | 151 struct drm_i915_private *dev_priv = ggtt->vm.i915; 161 ggtt->vm.reserved -= node->size; 174 struct drm_i915_private *dev_priv = ggtt->vm.i915; 177 if (!intel_vgpu_active(ggtt->vm.i915)) 190 struct drm_i915_private *dev_priv = ggtt->vm.i915; 200 ret = i915_gem_gtt_reserve(&ggtt->vm, NULL, node, 204 ggtt->vm.reserved += size; 255 struct drm_i915_private *dev_priv = ggtt->vm.i915; 257 unsigned long ggtt_end = ggtt->vm.total; 263 if (!intel_vgpu_active(ggtt->vm [all...] |
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dpi.c | 33 struct videomode *vm; member in struct:exynos_dpi 69 if (ctx->vm) { 78 drm_display_mode_from_videomode(ctx->vm, mode); 167 struct videomode *vm; local 172 vm = devm_kzalloc(dev, sizeof(*ctx->vm), GFP_KERNEL); 173 if (!vm) 176 ret = of_get_videomode(dn, vm, 0); 178 devm_kfree(dev, vm); 182 ctx->vm [all...] |
/linux-master/arch/s390/kernel/ |
H A D | early.c | 110 if (!memcmp(vmms->vm[0].cpi, "\xd2\xe5\xd4", 3)) 112 else if (!memcmp(vmms->vm[0].cpi, "\xa9\x61\xe5\xd4", 4)) 137 struct sysinfo_3_2_2 *vm = (struct sysinfo_3_2_2 *)&sysinfo_page; local 150 if (stsi(vm, 3, 2, 2) == 0 && vm->count) { 151 EBCASC(vm->vm[0].cpi, sizeof(vm->vm[0].cpi)); 152 sprintf(hvstr, "%-16.16s", vm [all...] |
H A D | lgr.c | 44 } vm[VM_LEVEL_MAX]; member in struct:lgr_info 105 cpascii(lgr_info->vm[i].name, si->vm[i].name, 106 sizeof(si->vm[i].name)); 107 cpascii(lgr_info->vm[i].cpi, si->vm[i].cpi, 108 sizeof(si->vm[i].cpi));
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_tlb.c | 36 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), 41 i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE); 50 batch = i915_gem_object_create_internal(ce->vm->i915, 4096); 54 vma = i915_vma_instance(batch, ce->vm, NULL); 67 addr = igt_random_offset(prng, addr, min(ce->vm->total, BIT_ULL(48)), 168 ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags); 171 tlbinv(ce->vm, addr & -length, length); 230 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) 283 va = i915_vma_instance(A, &ppgtt->vm, NUL 32 pte_tlbinv(struct intel_context *ce, struct i915_vma *va, struct i915_vma *vb, u64 align, void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), u64 length, struct rnd_state *prng) argument 228 mem_tlbinv(struct intel_gt *gt, struct drm_i915_gem_object *(*create_fn)(struct intel_gt *), void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) argument 363 tlbinv_full(struct i915_address_space *vm, u64 addr, u64 length) argument [all...] |
/linux-master/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 190 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); local 198 xe_vm_lock(vm, false); 199 bo = xe_bo_create_user(xe, NULL, vm, 0x10000, 203 xe_vm_unlock(vm); 275 down_read(&vm->lock); 276 xe_vm_lock(vm, false); 277 err = xe_bo_validate(bo, bo->vm, false); 278 xe_vm_unlock(vm); 279 up_read(&vm->lock); 321 xe_vm_put(vm); [all...] |
/linux-master/drivers/gpu/drm/arm/ |
H A D | hdlcd_crtc.c | 131 struct videomode vm; local 134 vm.vfront_porch = m->crtc_vsync_start - m->crtc_vdisplay; 135 vm.vback_porch = m->crtc_vtotal - m->crtc_vsync_end; 136 vm.vsync_len = m->crtc_vsync_end - m->crtc_vsync_start; 137 vm.hfront_porch = m->crtc_hsync_start - m->crtc_hdisplay; 138 vm.hback_porch = m->crtc_htotal - m->crtc_hsync_end; 139 vm.hsync_len = m->crtc_hsync_end - m->crtc_hsync_start; 153 hdlcd_write(hdlcd, HDLCD_REG_V_BACK_PORCH, vm.vback_porch - 1); 154 hdlcd_write(hdlcd, HDLCD_REG_V_FRONT_PORCH, vm.vfront_porch - 1); 155 hdlcd_write(hdlcd, HDLCD_REG_V_SYNC, vm [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_gem.c | 21 int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm) argument 89 if (vm) { 90 ret = lima_vm_map_bo(vm, bo, old_size >> PAGE_SHIFT); 167 struct lima_vm *vm = priv->vm; local 169 return lima_vm_bo_add(vm, bo, true); 176 struct lima_vm *vm = priv->vm; local 178 lima_vm_bo_del(vm, bo); 246 struct lima_vm *vm local 302 struct lima_vm *vm = priv->vm; local [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | run_vmtests.sh | 161 nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages) 166 echo 3 > /proc/sys/vm/drop_caches 167 if ! echo $((lackpgs + nr_hugepgs)) > /proc/sys/vm/nr_hugepages; then 216 echo 3 > /proc/sys/vm/drop_caches 218 echo 1 > /proc/sys/vm/compact_memory 265 nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages) 267 echo 1 > /proc/sys/vm/nr_hugepages 271 echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages 304 echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages 338 prev_policy=$(cat /proc/sys/vm/overcommit_memor [all...] |
/linux-master/drivers/gpu/drm/omapdrm/dss/ |
H A D | hdmi5_core.c | 239 video_cfg->v_fc_config.vm = cfg->vm; 243 video_cfg->hblank = cfg->vm.hfront_porch + 244 cfg->vm.hback_porch + cfg->vm.hsync_len; 246 video_cfg->vblank = cfg->vm.vsync_len + cfg->vm.vfront_porch + 247 cfg->vm.vback_porch; 250 if (cfg->vm.flags & DISPLAY_FLAGS_INTERLACED) { 255 video_cfg->v_fc_config.vm 276 const struct videomode *vm = &cfg->v_fc_config.vm; local 568 struct videomode vm; local [all...] |