/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | max_vcpuid_cap_test.c | 16 struct kvm_vm *vm; local 19 vm = vm_create_barebones(); 22 ret = vm_check_cap(vm, KVM_CAP_MAX_VCPU_ID); 25 ret = __vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, ret + 1); 30 vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, MAX_VCPU_ID); 34 ret = __vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, MAX_VCPU_ID + 1); 39 ret = __vm_ioctl(vm, KVM_CREATE_VCPU, (void *)MAX_VCPU_ID); 42 kvm_vm_free(vm);
|
H A D | nx_huge_pages_test.c | 72 static void check_2m_page_count(struct kvm_vm *vm, int expected_pages_2m) argument 76 actual_pages_2m = vm_get_stat(vm, "pages_2m"); 83 static void check_split_count(struct kvm_vm *vm, int expected_splits) argument 87 actual_splits = vm_get_stat(vm, "nx_lpage_splits"); 109 struct kvm_vm *vm; local 114 vm = vm_create(1); 117 r = __vm_disable_nx_huge_pages(vm); 127 vcpu = vm_vcpu_add(vm, 0, guest_code); 129 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS_HUGETLB, 133 nr_bytes = HPAGE_SLOT_NPAGES * vm [all...] |
H A D | exit_on_emulation_failure_test.c | 24 struct kvm_vm *vm; local 28 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 29 vm_enable_cap(vm, KVM_CAP_EXIT_ON_EMULATION_FAILURE, 1); 30 virt_map(vm, MMIO_GVA, MMIO_GPA, 1); 37 kvm_vm_free(vm);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, argument 35 static void mock_insert_entries(struct i915_address_space *vm, argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, argument 56 static void mock_cleanup(struct i915_address_space *vm) argument 60 static void mock_clear_range(struct i915_address_space *vm, argument 73 ppgtt->vm.gt = to_gt(i915); 74 ppgtt->vm.i915 = i915; 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); 76 ppgtt->vm 94 mock_bind_ggtt(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 102 mock_unbind_ggtt(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument [all...] |
/linux-master/samples/acrn/ |
H A D | Makefile | 3 .PHONY: vm-sample 5 vm-sample: vm-sample.o payload.o 12 rm *.o vm-sample
|
/linux-master/drivers/virt/acrn/ |
H A D | vm.c | 25 struct acrn_vm *acrn_vm_create(struct acrn_vm *vm, argument 37 mutex_init(&vm->regions_mapping_lock); 38 INIT_LIST_HEAD(&vm->ioreq_clients); 39 spin_lock_init(&vm->ioreq_clients_lock); 40 vm->vmid = vm_param->vmid; 41 vm->vcpu_num = vm_param->vcpu_num; 43 if (acrn_ioreq_init(vm, vm_param->ioreq_buf) < 0) { 45 vm->vmid = ACRN_INVALID_VMID; 50 list_add(&vm->list, &acrn_vm_list); 53 acrn_ioeventfd_init(vm); 59 acrn_vm_destroy(struct acrn_vm *vm) argument 104 acrn_msi_inject(struct acrn_vm *vm, u64 msi_addr, u64 msi_data) argument [all...] |
/linux-master/drivers/virtio/ |
H A D | virtio_mem.c | 283 static void virtio_mem_retry(struct virtio_mem *vm); 284 static int virtio_mem_create_resource(struct virtio_mem *vm); 285 static void virtio_mem_delete_resource(struct virtio_mem *vm); 291 static int register_virtio_mem_device(struct virtio_mem *vm) 300 list_add_rcu(&vm->next, &virtio_mem_devices); 310 static void unregister_virtio_mem_device(struct virtio_mem *vm) 314 list_del_rcu(&vm->next); 341 static unsigned long virtio_mem_phys_to_bb_id(struct virtio_mem *vm, 344 return addr / vm->bbm.bb_size; 350 static uint64_t virtio_mem_bb_id_to_phys(struct virtio_mem *vm, 287 register_virtio_mem_device(struct virtio_mem *vm) argument 306 unregister_virtio_mem_device(struct virtio_mem *vm) argument 337 virtio_mem_phys_to_bb_id(struct virtio_mem *vm, uint64_t addr) argument 346 virtio_mem_bb_id_to_phys(struct virtio_mem *vm, unsigned long bb_id) argument 355 virtio_mem_phys_to_sb_id(struct virtio_mem *vm, unsigned long addr) argument 367 virtio_mem_bbm_set_bb_state(struct virtio_mem *vm, unsigned long bb_id, enum virtio_mem_bbm_bb_state state) argument 385 virtio_mem_bbm_get_bb_state(struct virtio_mem *vm, unsigned long bb_id) argument 394 virtio_mem_bbm_bb_states_prepare_next_bb(struct virtio_mem *vm) argument 434 virtio_mem_sbm_set_mb_state(struct virtio_mem *vm, unsigned long mb_id, uint8_t state) argument 451 virtio_mem_sbm_get_mb_state(struct virtio_mem *vm, unsigned long mb_id) argument 462 virtio_mem_sbm_mb_states_prepare_next_mb(struct virtio_mem *vm) argument 501 virtio_mem_sbm_sb_state_bit_nr(struct virtio_mem *vm, unsigned long mb_id, int sb_id) argument 512 virtio_mem_sbm_set_sb_plugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 526 virtio_mem_sbm_set_sb_unplugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 538 virtio_mem_sbm_test_sb_plugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 555 virtio_mem_sbm_test_sb_unplugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 570 virtio_mem_sbm_first_unplugged_sb(struct virtio_mem *vm, unsigned long mb_id) argument 582 virtio_mem_sbm_sb_states_prepare_next_mb(struct virtio_mem *vm) argument 614 virtio_mem_could_add_memory(struct virtio_mem *vm, uint64_t size) argument 630 virtio_mem_add_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 666 virtio_mem_sbm_add_mb(struct virtio_mem *vm, unsigned long mb_id) argument 677 virtio_mem_bbm_add_bb(struct virtio_mem *vm, unsigned long bb_id) argument 694 virtio_mem_remove_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 718 virtio_mem_sbm_remove_mb(struct virtio_mem *vm, unsigned long mb_id) argument 734 virtio_mem_offline_and_remove_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 767 virtio_mem_sbm_offline_and_remove_mb(struct virtio_mem *vm, unsigned long mb_id) argument 782 virtio_mem_sbm_try_remove_unplugged_mb(struct virtio_mem *vm, unsigned long mb_id) argument 808 virtio_mem_bbm_offline_and_remove_bb(struct virtio_mem *vm, unsigned long bb_id) argument 820 virtio_mem_retry(struct virtio_mem *vm) argument 830 virtio_mem_translate_node_id(struct virtio_mem *vm, uint16_t node_id) argument 845 virtio_mem_overlaps_range(struct virtio_mem *vm, uint64_t start, uint64_t size) argument 855 virtio_mem_contains_range(struct virtio_mem *vm, uint64_t start, uint64_t size) argument 861 virtio_mem_sbm_notify_going_online(struct virtio_mem *vm, unsigned long mb_id) argument 876 virtio_mem_sbm_notify_offline(struct virtio_mem *vm, unsigned long mb_id) argument 896 virtio_mem_sbm_notify_online(struct virtio_mem *vm, unsigned long mb_id, unsigned long start_pfn) argument 921 virtio_mem_sbm_notify_going_offline(struct virtio_mem *vm, unsigned long mb_id) argument 937 virtio_mem_sbm_notify_cancel_offline(struct virtio_mem *vm, unsigned long mb_id) argument 953 virtio_mem_bbm_notify_going_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 968 virtio_mem_bbm_notify_cancel_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 987 struct virtio_mem *vm = container_of(nb, struct virtio_mem, local 1193 virtio_mem_fake_offline(struct virtio_mem *vm, unsigned long pfn, unsigned long nr_pages) argument 1275 virtio_mem_online_page(struct virtio_mem *vm, struct page *page, unsigned int order) argument 1337 struct virtio_mem *vm; local 1365 virtio_mem_send_request(struct virtio_mem *vm, const struct virtio_mem_req *req) argument 1395 virtio_mem_send_plug_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 1433 virtio_mem_send_unplug_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 1468 virtio_mem_send_unplug_all_request(struct virtio_mem *vm) argument 1499 virtio_mem_sbm_plug_sb(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 1517 virtio_mem_sbm_unplug_sb(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 1536 virtio_mem_bbm_unplug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1549 virtio_mem_bbm_plug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1566 virtio_mem_sbm_unplug_any_sb_raw(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1605 virtio_mem_sbm_unplug_mb(struct virtio_mem *vm, unsigned long mb_id) argument 1615 virtio_mem_sbm_prepare_next_mb(struct virtio_mem *vm, unsigned long *mb_id) argument 1644 virtio_mem_sbm_plug_and_add_mb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1695 virtio_mem_sbm_plug_any_sb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1736 virtio_mem_sbm_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1804 virtio_mem_bbm_plug_and_add_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1835 virtio_mem_bbm_prepare_next_bb(struct virtio_mem *vm, unsigned long *bb_id) argument 1854 virtio_mem_bbm_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1898 virtio_mem_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1914 virtio_mem_sbm_unplug_any_sb_offline(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1952 virtio_mem_sbm_unplug_sb_online(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 2000 virtio_mem_sbm_unplug_any_sb_online(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 2055 virtio_mem_sbm_unplug_any_sb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 2074 virtio_mem_sbm_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2134 virtio_mem_bbm_offline_remove_and_unplug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 2200 virtio_mem_bbm_bb_is_offline(struct virtio_mem *vm, unsigned long bb_id) argument 2219 virtio_mem_bbm_bb_is_movable(struct virtio_mem *vm, unsigned long bb_id) argument 2239 virtio_mem_bbm_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2282 virtio_mem_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2294 virtio_mem_cleanup_pending_mb(struct virtio_mem *vm) argument 2346 virtio_mem_refresh_config(struct virtio_mem *vm) argument 2393 struct virtio_mem *vm = container_of(work, struct virtio_mem, wq); local 2482 struct virtio_mem *vm = container_of(timer, struct virtio_mem, local 2493 struct virtio_mem *vm = vq->vdev->priv; local 2498 virtio_mem_init_vq(struct virtio_mem *vm) argument 2511 virtio_mem_init_hotplug(struct virtio_mem *vm) argument 2634 virtio_mem_send_state_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 2665 struct virtio_mem *vm = container_of(cb, struct virtio_mem, local 2700 virtio_mem_init_kdump(struct virtio_mem *vm) argument 2713 virtio_mem_init(struct virtio_mem *vm) argument 2754 virtio_mem_create_resource(struct virtio_mem *vm) argument 2782 virtio_mem_delete_resource(struct virtio_mem *vm) argument 2801 virtio_mem_has_memory_added(struct virtio_mem *vm) argument 2812 struct virtio_mem *vm; local 2861 virtio_mem_deinit_hotplug(struct virtio_mem *vm) argument 2926 virtio_mem_deinit_kdump(struct virtio_mem *vm) argument 2935 struct virtio_mem *vm = vdev->priv; local 2952 struct virtio_mem *vm = vdev->priv; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | sev.c | 17 static void encrypt_region(struct kvm_vm *vm, struct userspace_mem_region *region) argument 21 const sparsebit_idx_t lowest_page_in_region = gpa_base >> vm->page_shift; 27 sev_register_encrypted_memory(vm, region); 30 const uint64_t size = (j - i + 1) * vm->page_size; 31 const uint64_t offset = (i - lowest_page_in_region) * vm->page_size; 33 sev_launch_update_data(vm, gpa_base + offset, size); 37 void sev_vm_init(struct kvm_vm *vm) argument 39 if (vm->type == KVM_X86_DEFAULT_VM) { 40 assert(vm->arch.sev_fd == -1); 41 vm 50 sev_es_vm_init(struct kvm_vm *vm) argument 63 sev_vm_launch(struct kvm_vm *vm, uint32_t policy) argument 87 sev_vm_launch_measure(struct kvm_vm *vm, uint8_t *measurement) argument 100 sev_vm_launch_finish(struct kvm_vm *vm) argument 122 struct kvm_vm *vm; local 131 vm_sev_launch(struct kvm_vm *vm, uint32_t policy, uint8_t *measurement) argument [all...] |
H A D | hyperv.c | 11 struct hyperv_test_pages *vcpu_alloc_hyperv_test_pages(struct kvm_vm *vm, argument 14 vm_vaddr_t hv_pages_gva = vm_vaddr_alloc_page(vm); 15 struct hyperv_test_pages *hv = addr_gva2hva(vm, hv_pages_gva); 18 hv->vp_assist = (void *)vm_vaddr_alloc_page(vm); 19 hv->vp_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->vp_assist); 20 hv->vp_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->vp_assist); 23 hv->partition_assist = (void *)vm_vaddr_alloc_page(vm); 24 hv->partition_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->partition_assist); 25 hv->partition_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->partition_assist); 28 hv->enlightened_vmcs = (void *)vm_vaddr_alloc_page(vm); [all...] |
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | ucall.c | 11 void ucall_arch_init(struct kvm_vm *vm, vm_paddr_t mmio_gpa) argument 13 vm_vaddr_t mmio_gva = vm_vaddr_unused_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); 15 virt_map(vm, mmio_gva, mmio_gpa, 1); 17 vm->ucall_mmio_addr = mmio_gpa; 19 write_guest_global(vm, ucall_exit_mmio_addr, (vm_vaddr_t *)mmio_gva); 27 run->mmio.phys_addr == vcpu->vm->ucall_mmio_addr) {
|
H A D | processor.c | 23 static uint64_t page_align(struct kvm_vm *vm, uint64_t v) argument 25 return (v + vm->page_size) & ~(vm->page_size - 1); 28 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 30 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; 31 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; 36 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 38 unsigned int shift = 2 * (vm 47 pmd_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 58 pte_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 64 use_lpa2_pte_format(struct kvm_vm *vm) argument 70 addr_pte(struct kvm_vm *vm, uint64_t pa, uint64_t attrs) argument 88 pte_addr(struct kvm_vm *vm, uint64_t pte) argument 104 ptrs_per_pgd(struct kvm_vm *vm) argument 110 ptrs_per_pte(struct kvm_vm *vm) argument 115 virt_arch_pgd_alloc(struct kvm_vm *vm) argument 128 _virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr, uint64_t flags) argument 173 virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr) argument 180 virt_get_pte_hva(struct kvm_vm *vm, vm_vaddr_t gva) argument 218 addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) argument 225 pte_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t page, int level) argument 244 virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) argument 264 struct kvm_vm *vm = vcpu->vm; local 375 __aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, struct kvm_vcpu_init *init) argument 394 aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, struct kvm_vcpu_init *init, void *guest_code) argument 404 vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id) argument 496 vm_init_descriptor_tables(struct kvm_vm *vm) argument 504 vm_install_sync_handler(struct kvm_vm *vm, int vector, int ec, void (*handler)(struct ex_regs *)) argument 515 vm_install_exception_handler(struct kvm_vm *vm, int vector, void (*handler)(struct ex_regs *)) argument 633 vm_vaddr_populate_bitmap(struct kvm_vm *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/s390x/ |
H A D | cmma_test.c | 96 static void create_main_memslot(struct kvm_vm *vm) argument 100 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); 103 vm->memslots[i] = 0; 106 static void create_test_memslot(struct kvm_vm *vm) argument 108 vm_userspace_mem_region_add(vm, 110 TEST_DATA_START_GFN << vm->page_shift, 115 vm->memslots[MEM_REGION_TEST_DATA] = TEST_DATA_MEMSLOT; 118 static void create_memslots(struct kvm_vm *vm) argument 134 create_main_memslot(vm); 135 create_test_memslot(vm); 138 finish_vm_setup(struct kvm_vm *vm) argument 152 struct kvm_vm *vm; local 163 enable_cmma(struct kvm_vm *vm) argument 171 enable_dirty_tracking(struct kvm_vm *vm) argument 177 __enable_migration_mode(struct kvm_vm *vm) argument 186 enable_migration_mode(struct kvm_vm *vm) argument 193 is_migration_mode_on(struct kvm_vm *vm) argument 207 vm_get_cmma_bits(struct kvm_vm *vm, u64 flags, int *errno_out) argument 228 struct kvm_vm *vm = create_vm_two_memslots(); local 273 struct kvm_vm *vm = vm_create_barebones(); local 375 assert_all_slots_cmma_dirty(struct kvm_vm *vm) argument 425 assert_no_pages_cmma_dirty(struct kvm_vm *vm) argument 448 struct kvm_vm *vm = create_vm_two_memslots(); local 472 query_cmma_range(struct kvm_vm *vm, u64 start_gfn, u64 gfn_count, struct kvm_s390_cmma_log *res_out) argument 505 struct kvm_vm *vm = create_vm_two_memslots(); local 667 struct kvm_vm *vm = vm_create_barebones(); local [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_vm.c | 18 struct lima_vm *vm; member in struct:lima_bo_va 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) argument 43 vm->bts[pbe].cpu[bte] = 0; 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) argument 52 if (!vm->bts[pbe].cpu) { 57 vm->bts[pbe].cpu = dma_alloc_wc( 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); 60 if (!vm->bts[pbe].cpu) 63 pts = vm 77 lima_vm_bo_find(struct lima_vm *vm, struct lima_bo *bo) argument 91 lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create) argument 155 lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo) argument 185 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo) argument 202 struct lima_vm *vm; local 237 struct lima_vm *vm = container_of(kref, struct lima_vm, refcount); local 254 lima_vm_print(struct lima_vm *vm) argument 283 lima_vm_map_bo(struct lima_vm *vm, struct lima_bo *bo, int pageoff) argument [all...] |
H A D | lima_vm.h | 41 int lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create); 42 void lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo); 44 u32 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo); 49 static inline struct lima_vm *lima_vm_get(struct lima_vm *vm) argument 51 kref_get(&vm->refcount); 52 return vm; 55 static inline void lima_vm_put(struct lima_vm *vm) argument 57 if (vm) 58 kref_put(&vm->refcount, lima_vm_release); 61 void lima_vm_print(struct lima_vm *vm); [all...] |
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | sev.h | 30 void sev_vm_launch(struct kvm_vm *vm, uint32_t policy); 31 void sev_vm_launch_measure(struct kvm_vm *vm, uint8_t *measurement); 32 void sev_vm_launch_finish(struct kvm_vm *vm); 36 void vm_sev_launch(struct kvm_vm *vm, uint32_t policy, uint8_t *measurement); 47 #define __vm_sev_ioctl(vm, cmd, arg) \ 57 .sev_fd = (vm)->arch.sev_fd, \ 60 r = __vm_ioctl(vm, KVM_MEMORY_ENCRYPT_OP, &sev_cmd.raw); \ 64 #define vm_sev_ioctl(vm, cmd, arg) \ 66 int ret = __vm_sev_ioctl(vm, cmd, arg); \ 68 __TEST_ASSERT_VM_VCPU_IOCTL(!ret, #cmd, ret, vm); \ 74 sev_register_encrypted_memory(struct kvm_vm *vm, struct userspace_mem_region *region) argument 85 sev_launch_update_data(struct kvm_vm *vm, vm_paddr_t gpa, uint64_t size) argument [all...] |
H A D | ucall.h | 9 static inline void ucall_arch_init(struct kvm_vm *vm, vm_paddr_t mmio_gpa) argument
|
/linux-master/drivers/video/ |
H A D | videomode.c | 14 struct videomode *vm) 16 vm->pixelclock = dt->pixelclock.typ; 17 vm->hactive = dt->hactive.typ; 18 vm->hfront_porch = dt->hfront_porch.typ; 19 vm->hback_porch = dt->hback_porch.typ; 20 vm->hsync_len = dt->hsync_len.typ; 22 vm->vactive = dt->vactive.typ; 23 vm->vfront_porch = dt->vfront_porch.typ; 24 vm->vback_porch = dt->vback_porch.typ; 25 vm 13 videomode_from_timing(const struct display_timing *dt, struct videomode *vm) argument 31 videomode_from_timings(const struct display_timings *disp, struct videomode *vm, unsigned int index) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 31 static inline struct xe_vm *xe_vm_get(struct xe_vm *vm) argument 33 drm_gpuvm_get(&vm->gpuvm); 34 return vm; 37 static inline void xe_vm_put(struct xe_vm *vm) argument 39 drm_gpuvm_put(&vm->gpuvm); 42 int xe_vm_lock(struct xe_vm *vm, bool intr); 44 void xe_vm_unlock(struct xe_vm *vm); 46 static inline bool xe_vm_is_closed(struct xe_vm *vm) argument 48 /* Only guaranteed not to change when vm->lock is held */ 49 return !vm 52 xe_vm_is_banned(struct xe_vm *vm) argument 57 xe_vm_is_closed_or_banned(struct xe_vm *vm) argument 72 xe_vm_has_scratch(const struct xe_vm *vm) argument 186 xe_vm_in_fault_mode(struct xe_vm *vm) argument 191 xe_vm_in_lr_mode(struct xe_vm *vm) argument 196 xe_vm_in_preempt_fence_mode(struct xe_vm *vm) argument 214 xe_vm_queue_rebind_worker(struct xe_vm *vm) argument 229 xe_vm_reactivate_rebind(struct xe_vm *vm) argument 256 xe_vm_resv(struct xe_vm *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/lib/s390x/ |
H A D | processor.c | 13 void virt_arch_pgd_alloc(struct kvm_vm *vm) argument 17 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", 18 vm->page_size); 20 if (vm->pgd_created) 23 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, 25 vm->memslots[MEM_REGION_PT]); 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); 28 vm->pgd = paddr; 29 vm 37 virt_alloc_region(struct kvm_vm *vm, int ri) argument 50 virt_arch_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa) argument 89 addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) argument 114 virt_dump_ptes(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t ptea_start) argument 128 virt_dump_region(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t reg_tab_addr) argument 150 virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) argument 163 vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id) argument [all...] |
/linux-master/include/video/ |
H A D | of_videomode.h | 14 int of_get_videomode(struct device_node *np, struct videomode *vm,
|
/linux-master/sound/pci/ctxfi/ |
H A D | ctvmem.c | 26 * Find or create vm block based on requested @size. 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) argument 36 if (size > vm->size) { 42 mutex_lock(&vm->lock); 43 list_for_each(pos, &vm->unused) { 48 if (pos == &vm->unused) 52 /* Move the vm node from unused list to used list directly */ 53 list_move(&entry->list, &vm->used); 54 vm->size -= size; 65 list_add(&block->list, &vm 75 put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) argument 124 ct_vm_map(struct ct_vm *vm, struct snd_pcm_substream *substream, int size) argument 152 ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block) argument 164 ct_get_ptp_phys(struct ct_vm *vm, int index) argument 171 struct ct_vm *vm; local 215 ct_vm_destroy(struct ct_vm *vm) argument [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.h | 15 void intel_dpt_destroy(struct i915_address_space *vm); 16 struct i915_vma *intel_dpt_pin(struct i915_address_space *vm); 17 void intel_dpt_unpin(struct i915_address_space *vm);
|
/linux-master/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.h | 25 int panthor_vm_map_bo_range(struct panthor_vm *vm, struct panthor_gem_object *bo, 27 int panthor_vm_unmap_range(struct panthor_vm *vm, u64 va, u64 size); 29 panthor_vm_get_bo_for_va(struct panthor_vm *vm, u64 va, u64 *bo_offset); 31 int panthor_vm_active(struct panthor_vm *vm); 32 void panthor_vm_idle(struct panthor_vm *vm); 33 int panthor_vm_as(struct panthor_vm *vm); 36 panthor_vm_get_heap_pool(struct panthor_vm *vm, bool create); 38 struct panthor_vm *panthor_vm_get(struct panthor_vm *vm); 39 void panthor_vm_put(struct panthor_vm *vm); 46 struct panthor_vm *vm, [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_ggtt.c | 58 struct drm_i915_private *i915 = ggtt->vm.i915; 60 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); 62 ggtt->vm.is_ggtt = true; 65 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); 68 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; 74 ggtt->vm.cleanup(&ggtt->vm); 110 * @vm: The VM to suspend the mappings for 115 void i915_ggtt_suspend_vm(struct i915_address_space *vm) argument 120 drm_WARN_ON(&vm 436 gen8_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 451 gen8_ggtt_insert_page_bind(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 466 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 505 __gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 536 gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 549 gen8_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 570 gen8_ggtt_scratch_range_bind(struct i915_address_space *vm, u64 start, u64 length) argument 591 gen6_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 612 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 645 nop_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 650 bxt_vtd_ggtt_wa(struct i915_address_space *vm) argument 663 struct i915_address_space *vm; member in struct:insert_page 680 bxt_vtd_ggtt_insert_page__BKL(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 unused) argument 692 struct i915_address_space *vm; member in struct:insert_entries 709 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 719 gen6_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 740 intel_ggtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 764 intel_ggtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 917 aliasing_gtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 940 aliasing_gtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 1200 gen6_gmch_remove(struct i915_address_space *vm) argument 1556 i915_ggtt_resume_vm(struct i915_address_space *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | vcpu_width_config.c | 25 struct kvm_vm *vm; local 28 vm = vm_create_barebones(); 30 vcpu0 = __vm_vcpu_add(vm, 0); 35 vcpu1 = __vm_vcpu_add(vm, 1); 39 kvm_vm_free(vm); 51 struct kvm_vm *vm; local 54 vm = vm_create_barebones(); 56 vcpu0 = __vm_vcpu_add(vm, 0); 57 vcpu1 = __vm_vcpu_add(vm, 1); 66 kvm_vm_free(vm); 82 struct kvm_vm *vm; local [all...] |