/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | max_vcpuid_cap_test.c | 16 struct kvm_vm *vm; local 19 vm = vm_create_barebones(); 22 ret = vm_check_cap(vm, KVM_CAP_MAX_VCPU_ID); 25 ret = __vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, ret + 1); 30 vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, MAX_VCPU_ID); 34 ret = __vm_enable_cap(vm, KVM_CAP_MAX_VCPU_ID, MAX_VCPU_ID + 1); 39 ret = __vm_ioctl(vm, KVM_CREATE_VCPU, (void *)MAX_VCPU_ID); 42 kvm_vm_free(vm);
|
H A D | nx_huge_pages_test.c | 75 static void check_2m_page_count(struct kvm_vm *vm, int expected_pages_2m) argument 79 actual_pages_2m = vm_get_stat(vm, "pages_2m"); 86 static void check_split_count(struct kvm_vm *vm, int expected_splits) argument 90 actual_splits = vm_get_stat(vm, "nx_lpage_splits"); 112 struct kvm_vm *vm; local 117 vm = vm_create(1); 120 r = __vm_disable_nx_huge_pages(vm); 130 vcpu = vm_vcpu_add(vm, 0, guest_code); 132 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS_HUGETLB, 136 nr_bytes = HPAGE_SLOT_NPAGES * vm [all...] |
H A D | exit_on_emulation_failure_test.c | 27 struct kvm_vm *vm; local 31 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 32 vm_enable_cap(vm, KVM_CAP_EXIT_ON_EMULATION_FAILURE, 1); 33 virt_map(vm, MMIO_GVA, MMIO_GPA, 1); 40 kvm_vm_free(vm);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, argument 35 static void mock_insert_entries(struct i915_address_space *vm, argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, argument 56 static void mock_cleanup(struct i915_address_space *vm) argument 60 static void mock_clear_range(struct i915_address_space *vm, argument 73 ppgtt->vm.gt = to_gt(i915); 74 ppgtt->vm.i915 = i915; 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); 76 ppgtt->vm 94 mock_bind_ggtt(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 102 mock_unbind_ggtt(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument [all...] |
/linux-master/samples/acrn/ |
H A D | Makefile | 3 .PHONY: vm-sample 5 vm-sample: vm-sample.o payload.o 12 rm *.o vm-sample
|
/linux-master/drivers/virt/acrn/ |
H A D | vm.c | 25 struct acrn_vm *acrn_vm_create(struct acrn_vm *vm, argument 37 mutex_init(&vm->regions_mapping_lock); 38 INIT_LIST_HEAD(&vm->ioreq_clients); 39 spin_lock_init(&vm->ioreq_clients_lock); 40 vm->vmid = vm_param->vmid; 41 vm->vcpu_num = vm_param->vcpu_num; 43 if (acrn_ioreq_init(vm, vm_param->ioreq_buf) < 0) { 45 vm->vmid = ACRN_INVALID_VMID; 50 list_add(&vm->list, &acrn_vm_list); 53 acrn_ioeventfd_init(vm); 59 acrn_vm_destroy(struct acrn_vm *vm) argument 104 acrn_msi_inject(struct acrn_vm *vm, u64 msi_addr, u64 msi_data) argument [all...] |
H A D | irqfd.c | 23 * @vm: Associated VM pointer 32 struct acrn_vm *vm; member in struct:hsm_irqfd 43 struct acrn_vm *vm = irqfd->vm; local 45 acrn_msi_inject(vm, irqfd->msi.msi_addr, 53 lockdep_assert_held(&irqfd->vm->irqfds_lock); 65 struct acrn_vm *vm; local 68 vm = irqfd->vm; 69 mutex_lock(&vm 81 struct acrn_vm *vm; local 110 acrn_irqfd_assign(struct acrn_vm *vm, struct acrn_irqfd *args) argument 177 acrn_irqfd_deassign(struct acrn_vm *vm, struct acrn_irqfd *args) argument 200 acrn_irqfd_config(struct acrn_vm *vm, struct acrn_irqfd *args) argument 212 acrn_irqfd_init(struct acrn_vm *vm) argument 224 acrn_irqfd_deinit(struct acrn_vm *vm) argument [all...] |
/linux-master/drivers/virtio/ |
H A D | virtio_mem.c | 278 static void virtio_mem_retry(struct virtio_mem *vm); 279 static int virtio_mem_create_resource(struct virtio_mem *vm); 280 static void virtio_mem_delete_resource(struct virtio_mem *vm); 286 static int register_virtio_mem_device(struct virtio_mem *vm) argument 295 list_add_rcu(&vm->next, &virtio_mem_devices); 305 static void unregister_virtio_mem_device(struct virtio_mem *vm) argument 309 list_del_rcu(&vm->next); 336 static unsigned long virtio_mem_phys_to_bb_id(struct virtio_mem *vm, argument 339 return addr / vm->bbm.bb_size; 345 static uint64_t virtio_mem_bb_id_to_phys(struct virtio_mem *vm, argument 354 virtio_mem_phys_to_sb_id(struct virtio_mem *vm, unsigned long addr) argument 366 virtio_mem_bbm_set_bb_state(struct virtio_mem *vm, unsigned long bb_id, enum virtio_mem_bbm_bb_state state) argument 384 virtio_mem_bbm_get_bb_state(struct virtio_mem *vm, unsigned long bb_id) argument 393 virtio_mem_bbm_bb_states_prepare_next_bb(struct virtio_mem *vm) argument 433 virtio_mem_sbm_set_mb_state(struct virtio_mem *vm, unsigned long mb_id, uint8_t state) argument 450 virtio_mem_sbm_get_mb_state(struct virtio_mem *vm, unsigned long mb_id) argument 461 virtio_mem_sbm_mb_states_prepare_next_mb(struct virtio_mem *vm) argument 500 virtio_mem_sbm_sb_state_bit_nr(struct virtio_mem *vm, unsigned long mb_id, int sb_id) argument 511 virtio_mem_sbm_set_sb_plugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 525 virtio_mem_sbm_set_sb_unplugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 537 virtio_mem_sbm_test_sb_plugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 554 virtio_mem_sbm_test_sb_unplugged(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 569 virtio_mem_sbm_first_unplugged_sb(struct virtio_mem *vm, unsigned long mb_id) argument 581 virtio_mem_sbm_sb_states_prepare_next_mb(struct virtio_mem *vm) argument 613 virtio_mem_could_add_memory(struct virtio_mem *vm, uint64_t size) argument 629 virtio_mem_add_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 665 virtio_mem_sbm_add_mb(struct virtio_mem *vm, unsigned long mb_id) argument 676 virtio_mem_bbm_add_bb(struct virtio_mem *vm, unsigned long bb_id) argument 693 virtio_mem_remove_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 717 virtio_mem_sbm_remove_mb(struct virtio_mem *vm, unsigned long mb_id) argument 733 virtio_mem_offline_and_remove_memory(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 766 virtio_mem_sbm_offline_and_remove_mb(struct virtio_mem *vm, unsigned long mb_id) argument 781 virtio_mem_sbm_try_remove_unplugged_mb(struct virtio_mem *vm, unsigned long mb_id) argument 807 virtio_mem_bbm_offline_and_remove_bb(struct virtio_mem *vm, unsigned long bb_id) argument 819 virtio_mem_retry(struct virtio_mem *vm) argument 829 virtio_mem_translate_node_id(struct virtio_mem *vm, uint16_t node_id) argument 844 virtio_mem_overlaps_range(struct virtio_mem *vm, uint64_t start, uint64_t size) argument 854 virtio_mem_contains_range(struct virtio_mem *vm, uint64_t start, uint64_t size) argument 860 virtio_mem_sbm_notify_going_online(struct virtio_mem *vm, unsigned long mb_id) argument 875 virtio_mem_sbm_notify_offline(struct virtio_mem *vm, unsigned long mb_id) argument 895 virtio_mem_sbm_notify_online(struct virtio_mem *vm, unsigned long mb_id, unsigned long start_pfn) argument 920 virtio_mem_sbm_notify_going_offline(struct virtio_mem *vm, unsigned long mb_id) argument 936 virtio_mem_sbm_notify_cancel_offline(struct virtio_mem *vm, unsigned long mb_id) argument 952 virtio_mem_bbm_notify_going_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 967 virtio_mem_bbm_notify_cancel_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 986 struct virtio_mem *vm = container_of(nb, struct virtio_mem, local 1192 virtio_mem_fake_offline(struct virtio_mem *vm, unsigned long pfn, unsigned long nr_pages) argument 1274 virtio_mem_online_page(struct virtio_mem *vm, struct page *page, unsigned int order) argument 1336 struct virtio_mem *vm; local 1364 virtio_mem_send_request(struct virtio_mem *vm, const struct virtio_mem_req *req) argument 1394 virtio_mem_send_plug_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 1432 virtio_mem_send_unplug_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 1467 virtio_mem_send_unplug_all_request(struct virtio_mem *vm) argument 1498 virtio_mem_sbm_plug_sb(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 1516 virtio_mem_sbm_unplug_sb(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 1535 virtio_mem_bbm_unplug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1548 virtio_mem_bbm_plug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1565 virtio_mem_sbm_unplug_any_sb_raw(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1604 virtio_mem_sbm_unplug_mb(struct virtio_mem *vm, unsigned long mb_id) argument 1614 virtio_mem_sbm_prepare_next_mb(struct virtio_mem *vm, unsigned long *mb_id) argument 1643 virtio_mem_sbm_plug_and_add_mb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1694 virtio_mem_sbm_plug_any_sb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1735 virtio_mem_sbm_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1803 virtio_mem_bbm_plug_and_add_bb(struct virtio_mem *vm, unsigned long bb_id) argument 1834 virtio_mem_bbm_prepare_next_bb(struct virtio_mem *vm, unsigned long *bb_id) argument 1853 virtio_mem_bbm_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1897 virtio_mem_plug_request(struct virtio_mem *vm, uint64_t diff) argument 1913 virtio_mem_sbm_unplug_any_sb_offline(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 1951 virtio_mem_sbm_unplug_sb_online(struct virtio_mem *vm, unsigned long mb_id, int sb_id, int count) argument 1999 virtio_mem_sbm_unplug_any_sb_online(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 2054 virtio_mem_sbm_unplug_any_sb(struct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) argument 2073 virtio_mem_sbm_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2133 virtio_mem_bbm_offline_remove_and_unplug_bb(struct virtio_mem *vm, unsigned long bb_id) argument 2199 virtio_mem_bbm_bb_is_offline(struct virtio_mem *vm, unsigned long bb_id) argument 2218 virtio_mem_bbm_bb_is_movable(struct virtio_mem *vm, unsigned long bb_id) argument 2238 virtio_mem_bbm_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2281 virtio_mem_unplug_request(struct virtio_mem *vm, uint64_t diff) argument 2293 virtio_mem_cleanup_pending_mb(struct virtio_mem *vm) argument 2345 virtio_mem_refresh_config(struct virtio_mem *vm) argument 2392 struct virtio_mem *vm = container_of(work, struct virtio_mem, wq); local 2481 struct virtio_mem *vm = container_of(timer, struct virtio_mem, local 2492 struct virtio_mem *vm = vq->vdev->priv; local 2497 virtio_mem_init_vq(struct virtio_mem *vm) argument 2510 virtio_mem_init_hotplug(struct virtio_mem *vm) argument 2633 virtio_mem_send_state_request(struct virtio_mem *vm, uint64_t addr, uint64_t size) argument 2664 struct virtio_mem *vm = container_of(cb, struct virtio_mem, local 2699 virtio_mem_init_kdump(struct virtio_mem *vm) argument 2712 virtio_mem_init(struct virtio_mem *vm) argument 2753 virtio_mem_create_resource(struct virtio_mem *vm) argument 2781 virtio_mem_delete_resource(struct virtio_mem *vm) argument 2800 virtio_mem_has_memory_added(struct virtio_mem *vm) argument 2811 struct virtio_mem *vm; local 2860 virtio_mem_deinit_hotplug(struct virtio_mem *vm) argument 2925 virtio_mem_deinit_kdump(struct virtio_mem *vm) argument 2934 struct virtio_mem *vm = vdev->priv; local 2951 struct virtio_mem *vm = vdev->priv; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | ucall.c | 11 void ucall_arch_init(struct kvm_vm *vm, vm_paddr_t mmio_gpa) argument 13 vm_vaddr_t mmio_gva = vm_vaddr_unused_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); 15 virt_map(vm, mmio_gva, mmio_gpa, 1); 17 vm->ucall_mmio_addr = mmio_gpa; 19 write_guest_global(vm, ucall_exit_mmio_addr, (vm_vaddr_t *)mmio_gva); 27 run->mmio.phys_addr == vcpu->vm->ucall_mmio_addr) {
|
H A D | processor.c | 21 static uint64_t page_align(struct kvm_vm *vm, uint64_t v) argument 23 return (v + vm->page_size) & ~(vm->page_size - 1); 26 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 28 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; 29 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; 34 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 36 unsigned int shift = 2 * (vm 45 pmd_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 56 pte_index(struct kvm_vm *vm, vm_vaddr_t gva) argument 62 use_lpa2_pte_format(struct kvm_vm *vm) argument 68 addr_pte(struct kvm_vm *vm, uint64_t pa, uint64_t attrs) argument 86 pte_addr(struct kvm_vm *vm, uint64_t pte) argument 102 ptrs_per_pgd(struct kvm_vm *vm) argument 108 ptrs_per_pte(struct kvm_vm *vm) argument 113 virt_arch_pgd_alloc(struct kvm_vm *vm) argument 126 _virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr, uint64_t flags) argument 171 virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr) argument 178 virt_get_pte_hva(struct kvm_vm *vm, vm_vaddr_t gva) argument 216 addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) argument 223 pte_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t page, int level) argument 242 virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) argument 262 struct kvm_vm *vm = vcpu->vm; local 373 __aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, struct kvm_vcpu_init *init) argument 392 aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, struct kvm_vcpu_init *init, void *guest_code) argument 402 vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id) argument 494 vm_init_descriptor_tables(struct kvm_vm *vm) argument 502 vm_install_sync_handler(struct kvm_vm *vm, int vector, int ec, void (*handler)(struct ex_regs *)) argument 513 vm_install_exception_handler(struct kvm_vm *vm, int vector, void (*handler)(struct ex_regs *)) argument 631 vm_vaddr_populate_bitmap(struct kvm_vm *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | sev.h | 30 void sev_vm_launch(struct kvm_vm *vm, uint32_t policy); 31 void sev_vm_launch_measure(struct kvm_vm *vm, uint8_t *measurement); 32 void sev_vm_launch_finish(struct kvm_vm *vm); 46 #define __vm_sev_ioctl(vm, cmd, arg) \ 56 .sev_fd = (vm)->arch.sev_fd, \ 59 r = __vm_ioctl(vm, KVM_MEMORY_ENCRYPT_OP, &sev_cmd.raw); \ 63 #define vm_sev_ioctl(vm, cmd, arg) \ 65 int ret = __vm_sev_ioctl(vm, cmd, arg); \ 67 __TEST_ASSERT_VM_VCPU_IOCTL(!ret, #cmd, ret, vm); \ 70 static inline void sev_vm_init(struct kvm_vm *vm) argument 78 sev_es_vm_init(struct kvm_vm *vm) argument 85 sev_register_encrypted_memory(struct kvm_vm *vm, struct userspace_mem_region *region) argument 96 sev_launch_update_data(struct kvm_vm *vm, vm_paddr_t gpa, uint64_t size) argument [all...] |
H A D | ucall.h | 9 static inline void ucall_arch_init(struct kvm_vm *vm, vm_paddr_t mmio_gpa) argument
|
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | sev.c | 18 static void encrypt_region(struct kvm_vm *vm, struct userspace_mem_region *region) argument 22 const sparsebit_idx_t lowest_page_in_region = gpa_base >> vm->page_shift; 28 sev_register_encrypted_memory(vm, region); 31 const uint64_t size = (j - i + 1) * vm->page_size; 32 const uint64_t offset = (i - lowest_page_in_region) * vm->page_size; 34 sev_launch_update_data(vm, gpa_base + offset, size); 38 void sev_vm_launch(struct kvm_vm *vm, uint32_t policy) argument 47 vm_sev_ioctl(vm, KVM_SEV_LAUNCH_START, &launch_start); 48 vm_sev_ioctl(vm, KVM_SEV_GUEST_STATUS, &status); 53 hash_for_each(vm 62 sev_vm_launch_measure(struct kvm_vm *vm, uint8_t *measurement) argument 75 sev_vm_launch_finish(struct kvm_vm *vm) argument 99 struct kvm_vm *vm; local [all...] |
H A D | hyperv.c | 11 struct hyperv_test_pages *vcpu_alloc_hyperv_test_pages(struct kvm_vm *vm, argument 14 vm_vaddr_t hv_pages_gva = vm_vaddr_alloc_page(vm); 15 struct hyperv_test_pages *hv = addr_gva2hva(vm, hv_pages_gva); 18 hv->vp_assist = (void *)vm_vaddr_alloc_page(vm); 19 hv->vp_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->vp_assist); 20 hv->vp_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->vp_assist); 23 hv->partition_assist = (void *)vm_vaddr_alloc_page(vm); 24 hv->partition_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->partition_assist); 25 hv->partition_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->partition_assist); 28 hv->enlightened_vmcs = (void *)vm_vaddr_alloc_page(vm); [all...] |
/linux-master/tools/testing/selftests/kvm/s390x/ |
H A D | cmma_test.c | 97 static void create_main_memslot(struct kvm_vm *vm) argument 101 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); 104 vm->memslots[i] = 0; 107 static void create_test_memslot(struct kvm_vm *vm) argument 109 vm_userspace_mem_region_add(vm, 111 TEST_DATA_START_GFN << vm->page_shift, 116 vm->memslots[MEM_REGION_TEST_DATA] = TEST_DATA_MEMSLOT; 119 static void create_memslots(struct kvm_vm *vm) argument 135 create_main_memslot(vm); 136 create_test_memslot(vm); 139 finish_vm_setup(struct kvm_vm *vm) argument 153 struct kvm_vm *vm; local 164 enable_cmma(struct kvm_vm *vm) argument 172 enable_dirty_tracking(struct kvm_vm *vm) argument 178 __enable_migration_mode(struct kvm_vm *vm) argument 187 enable_migration_mode(struct kvm_vm *vm) argument 194 is_migration_mode_on(struct kvm_vm *vm) argument 208 vm_get_cmma_bits(struct kvm_vm *vm, u64 flags, int *errno_out) argument 229 struct kvm_vm *vm = create_vm_two_memslots(); local 274 struct kvm_vm *vm = vm_create_barebones(); local 376 assert_all_slots_cmma_dirty(struct kvm_vm *vm) argument 426 assert_no_pages_cmma_dirty(struct kvm_vm *vm) argument 449 struct kvm_vm *vm = create_vm_two_memslots(); local 473 query_cmma_range(struct kvm_vm *vm, u64 start_gfn, u64 gfn_count, struct kvm_s390_cmma_log *res_out) argument 506 struct kvm_vm *vm = create_vm_two_memslots(); local 668 struct kvm_vm *vm = vm_create_barebones(); local [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_vm.c | 18 struct lima_vm *vm; member in struct:lima_bo_va 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) argument 43 vm->bts[pbe].cpu[bte] = 0; 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) argument 52 if (!vm->bts[pbe].cpu) { 57 vm->bts[pbe].cpu = dma_alloc_wc( 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); 60 if (!vm->bts[pbe].cpu) 63 pts = vm 77 lima_vm_bo_find(struct lima_vm *vm, struct lima_bo *bo) argument 91 lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create) argument 155 lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo) argument 185 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo) argument 202 struct lima_vm *vm; local 237 struct lima_vm *vm = container_of(kref, struct lima_vm, refcount); local 254 lima_vm_print(struct lima_vm *vm) argument 283 lima_vm_map_bo(struct lima_vm *vm, struct lima_bo *bo, int pageoff) argument [all...] |
H A D | lima_vm.h | 41 int lima_vm_bo_add(struct lima_vm *vm, struct lima_bo *bo, bool create); 42 void lima_vm_bo_del(struct lima_vm *vm, struct lima_bo *bo); 44 u32 lima_vm_get_va(struct lima_vm *vm, struct lima_bo *bo); 49 static inline struct lima_vm *lima_vm_get(struct lima_vm *vm) argument 51 kref_get(&vm->refcount); 52 return vm; 55 static inline void lima_vm_put(struct lima_vm *vm) argument 57 if (vm) 58 kref_put(&vm->refcount, lima_vm_release); 61 void lima_vm_print(struct lima_vm *vm); [all...] |
/linux-master/drivers/video/ |
H A D | videomode.c | 14 struct videomode *vm) 16 vm->pixelclock = dt->pixelclock.typ; 17 vm->hactive = dt->hactive.typ; 18 vm->hfront_porch = dt->hfront_porch.typ; 19 vm->hback_porch = dt->hback_porch.typ; 20 vm->hsync_len = dt->hsync_len.typ; 22 vm->vactive = dt->vactive.typ; 23 vm->vfront_porch = dt->vfront_porch.typ; 24 vm->vback_porch = dt->vback_porch.typ; 25 vm 13 videomode_from_timing(const struct display_timing *dt, struct videomode *vm) argument 31 videomode_from_timings(const struct display_timings *disp, struct videomode *vm, unsigned int index) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 31 static inline struct xe_vm *xe_vm_get(struct xe_vm *vm) argument 33 drm_gpuvm_get(&vm->gpuvm); 34 return vm; 37 static inline void xe_vm_put(struct xe_vm *vm) argument 39 drm_gpuvm_put(&vm->gpuvm); 42 int xe_vm_lock(struct xe_vm *vm, bool intr); 44 void xe_vm_unlock(struct xe_vm *vm); 46 static inline bool xe_vm_is_closed(struct xe_vm *vm) argument 48 /* Only guaranteed not to change when vm->lock is held */ 49 return !vm 52 xe_vm_is_banned(struct xe_vm *vm) argument 57 xe_vm_is_closed_or_banned(struct xe_vm *vm) argument 72 xe_vm_has_scratch(const struct xe_vm *vm) argument 186 xe_vm_in_fault_mode(struct xe_vm *vm) argument 191 xe_vm_in_lr_mode(struct xe_vm *vm) argument 196 xe_vm_in_preempt_fence_mode(struct xe_vm *vm) argument 214 xe_vm_queue_rebind_worker(struct xe_vm *vm) argument 229 xe_vm_reactivate_rebind(struct xe_vm *vm) argument 256 xe_vm_resv(struct xe_vm *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/lib/s390x/ |
H A D | processor.c | 13 void virt_arch_pgd_alloc(struct kvm_vm *vm) argument 17 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", 18 vm->page_size); 20 if (vm->pgd_created) 23 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, 25 vm->memslots[MEM_REGION_PT]); 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); 28 vm->pgd = paddr; 29 vm 37 virt_alloc_region(struct kvm_vm *vm, int ri) argument 50 virt_arch_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa) argument 89 addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) argument 114 virt_dump_ptes(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t ptea_start) argument 128 virt_dump_region(FILE *stream, struct kvm_vm *vm, uint8_t indent, uint64_t reg_tab_addr) argument 150 virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) argument 163 vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id) argument [all...] |
/linux-master/include/video/ |
H A D | of_videomode.h | 14 int of_get_videomode(struct device_node *np, struct videomode *vm,
|
/linux-master/sound/pci/ctxfi/ |
H A D | ctvmem.c | 26 * Find or create vm block based on requested @size. 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) argument 36 if (size > vm->size) { 42 mutex_lock(&vm->lock); 43 list_for_each(pos, &vm->unused) { 48 if (pos == &vm->unused) 52 /* Move the vm node from unused list to used list directly */ 53 list_move(&entry->list, &vm->used); 54 vm->size -= size; 65 list_add(&block->list, &vm 75 put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) argument 124 ct_vm_map(struct ct_vm *vm, struct snd_pcm_substream *substream, int size) argument 152 ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block) argument 164 ct_get_ptp_phys(struct ct_vm *vm, int index) argument 171 struct ct_vm *vm; local 215 ct_vm_destroy(struct ct_vm *vm) argument [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.h | 15 void intel_dpt_destroy(struct i915_address_space *vm); 16 struct i915_vma *intel_dpt_pin(struct i915_address_space *vm); 17 void intel_dpt_unpin(struct i915_address_space *vm);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_ggtt.c | 58 struct drm_i915_private *i915 = ggtt->vm.i915; 60 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); 62 ggtt->vm.is_ggtt = true; 65 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); 68 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; 74 ggtt->vm.cleanup(&ggtt->vm); 110 * @vm: The VM to suspend the mappings for 115 void i915_ggtt_suspend_vm(struct i915_address_space *vm) argument 120 drm_WARN_ON(&vm 439 gen8_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 454 gen8_ggtt_insert_page_bind(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 469 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 508 __gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 539 gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 552 gen8_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 573 gen8_ggtt_scratch_range_bind(struct i915_address_space *vm, u64 start, u64 length) argument 594 gen6_ggtt_insert_page(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 flags) argument 615 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 648 nop_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 653 bxt_vtd_ggtt_wa(struct i915_address_space *vm) argument 666 struct i915_address_space *vm; member in struct:insert_page 683 bxt_vtd_ggtt_insert_page__BKL(struct i915_address_space *vm, dma_addr_t addr, u64 offset, unsigned int pat_index, u32 unused) argument 695 struct i915_address_space *vm; member in struct:insert_entries 712 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 722 gen6_ggtt_clear_range(struct i915_address_space *vm, u64 start, u64 length) argument 743 intel_ggtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 767 intel_ggtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 920 aliasing_gtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 943 aliasing_gtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 1203 gen6_gmch_remove(struct i915_address_space *vm) argument 1559 i915_ggtt_resume_vm(struct i915_address_space *vm) argument [all...] |
/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | vcpu_width_config.c | 25 struct kvm_vm *vm; local 28 vm = vm_create_barebones(); 30 vcpu0 = __vm_vcpu_add(vm, 0); 35 vcpu1 = __vm_vcpu_add(vm, 1); 39 kvm_vm_free(vm); 51 struct kvm_vm *vm; local 54 vm = vm_create_barebones(); 56 vcpu0 = __vm_vcpu_add(vm, 0); 57 vcpu1 = __vm_vcpu_add(vm, 1); 66 kvm_vm_free(vm); 82 struct kvm_vm *vm; local [all...] |