/linux-master/arch/x86/virt/ |
H A D | Makefile | 2 obj-y += svm/ vmx/
|
/linux-master/arch/x86/kvm/svm/ |
H A D | hyperv.c | 11 struct vcpu_svm *svm = to_svm(vcpu); local 13 svm->vmcb->control.exit_code = HV_SVM_EXITCODE_ENL; 14 svm->vmcb->control.exit_code_hi = 0; 15 svm->vmcb->control.exit_info_1 = HV_SVM_ENL_EXITCODE_TRAP_AFTER_FLUSH; 16 svm->vmcb->control.exit_info_2 = 0; 17 nested_svm_vmexit(svm);
|
H A D | nested.c | 31 #include "svm.h" 39 struct vcpu_svm *svm = to_svm(vcpu); local 40 struct vmcb *vmcb = svm->vmcb; 56 nested_svm_vmexit(svm); 61 struct vcpu_svm *svm = to_svm(vcpu); local 62 u64 cr3 = svm->nested.ctl.nested_cr3; 75 struct vcpu_svm *svm = to_svm(vcpu); local 77 return svm->nested.ctl.nested_cr3; 82 struct vcpu_svm *svm = to_svm(vcpu); local 93 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm 108 nested_vmcb_needs_vls_intercept(struct vcpu_svm *svm) argument 122 recalc_intercepts(struct vcpu_svm *svm) argument 188 nested_svm_vmrun_msrpm(struct vcpu_svm *svm) argument 321 struct vcpu_svm *svm = to_svm(vcpu); local 329 struct vcpu_svm *svm = to_svm(vcpu); local 382 nested_copy_vmcb_control_to_cache(struct vcpu_svm *svm, struct vmcb_control_area *control) argument 404 nested_copy_vmcb_save_to_cache(struct vcpu_svm *svm, struct vmcb_save_area *save) argument 414 nested_sync_control_from_vmcb02(struct vcpu_svm *svm) argument 448 nested_save_pending_event_to_vmcb12(struct vcpu_svm *svm, struct vmcb *vmcb12) argument 528 nested_vmcb02_compute_g_pat(struct vcpu_svm *svm) argument 537 nested_vmcb02_prepare_save(struct vcpu_svm *svm, struct vmcb *vmcb12) argument 632 nested_vmcb02_prepare_control(struct vcpu_svm *svm, unsigned long vmcb12_rip, unsigned long vmcb12_csbase) argument 788 struct vcpu_svm *svm = to_svm(vcpu); local 839 struct vcpu_svm *svm = to_svm(vcpu); local 967 nested_svm_vmexit(struct vcpu_svm *svm) argument 1168 struct vcpu_svm *svm = to_svm(vcpu); local 1177 svm_allocate_nested(struct vcpu_svm *svm) argument 1203 svm_free_nested(struct vcpu_svm *svm) argument 1231 struct vcpu_svm *svm = to_svm(vcpu); local 1251 nested_svm_exit_handled_msr(struct vcpu_svm *svm) argument 1276 nested_svm_intercept_ioio(struct vcpu_svm *svm) argument 1301 nested_svm_intercept(struct vcpu_svm *svm) argument 1345 nested_svm_exit_handled(struct vcpu_svm *svm) argument 1375 struct vcpu_svm *svm = to_svm(vcpu); local 1383 struct vcpu_svm *svm = to_svm(vcpu); local 1416 nested_exit_on_init(struct vcpu_svm *svm) argument 1424 struct vcpu_svm *svm = to_svm(vcpu); local 1496 nested_svm_exit_special(struct vcpu_svm *svm) argument 1534 struct vcpu_svm *svm = to_svm(vcpu); local 1582 struct vcpu_svm *svm; local 1649 struct vcpu_svm *svm = to_svm(vcpu); local 1782 struct vcpu_svm *svm = to_svm(vcpu); local [all...] |
H A D | svm.c | 49 #include "svm.h" 298 struct vcpu_svm *svm = to_svm(vcpu); local 313 svm_set_gif(svm, true); 316 clr_exception_intercept(svm, GP_VECTOR); 324 svm_free_nested(svm); 327 int ret = svm_allocate_nested(svm); 339 set_exception_intercept(svm, GP_VECTOR); 343 svm->vmcb->save.efer = efer | EFER_SVME; 344 vmcb_mark_dirty(svm->vmcb, VMCB_CR); 350 struct vcpu_svm *svm local 360 struct vcpu_svm *svm = to_svm(vcpu); local 372 struct vcpu_svm *svm = to_svm(vcpu); local 415 struct vcpu_svm *svm = to_svm(vcpu); local 460 struct vcpu_svm *svm = to_svm(vcpu); local 724 set_dr_intercepts(struct vcpu_svm *svm) argument 748 clr_dr_intercepts(struct vcpu_svm *svm) argument 771 struct vcpu_svm *svm = to_svm(vcpu); local 825 struct vcpu_svm *svm = to_svm(vcpu); local 892 svm_set_x2apic_msr_interception(struct vcpu_svm *svm, bool intercept) argument 922 struct vcpu_svm *svm = to_svm(vcpu); local 995 struct vcpu_svm *svm = to_svm(vcpu); local 1010 struct vcpu_svm *svm = to_svm(vcpu); local 1026 svm_get_lbr_vmcb(struct vcpu_svm *svm) argument 1039 struct vcpu_svm *svm = to_svm(vcpu); local 1054 disable_nmi_singlestep(struct vcpu_svm *svm) argument 1069 struct vcpu_svm *svm = to_svm(vcpu); local 1090 struct vcpu_svm *svm = to_svm(vcpu); local 1142 struct vcpu_svm *svm = to_svm(vcpu); local 1149 struct vcpu_svm *svm = to_svm(vcpu); local 1156 struct vcpu_svm *svm = to_svm(vcpu); local 1172 svm_recalc_instruction_intercepts(struct kvm_vcpu *vcpu, struct vcpu_svm *svm) argument 1197 struct vcpu_svm *svm = to_svm(vcpu); local 1229 struct vcpu_svm *svm = to_svm(vcpu); local 1379 struct vcpu_svm *svm = to_svm(vcpu); local 1396 struct vcpu_svm *svm = to_svm(vcpu); local 1407 svm_switch_vmcb(struct vcpu_svm *svm, struct kvm_vmcb_info *target_vmcb) argument 1415 struct vcpu_svm *svm; local 1488 struct vcpu_svm *svm = to_svm(vcpu); local 1513 struct vcpu_svm *svm = to_svm(vcpu); local 1553 struct vcpu_svm *svm = to_svm(vcpu); local 1578 struct vcpu_svm *svm = to_svm(vcpu); local 1631 svm_set_vintr(struct vcpu_svm *svm) argument 1664 svm_clear_vintr(struct vcpu_svm *svm) argument 1798 struct vcpu_svm *svm = to_svm(vcpu); local 1806 struct vcpu_svm *svm = to_svm(vcpu); local 1815 struct vcpu_svm *svm = to_svm(vcpu); local 1823 struct vcpu_svm *svm = to_svm(vcpu); local 1832 struct vcpu_svm *svm = to_svm(vcpu); local 1855 struct vcpu_svm *svm = to_svm(vcpu); local 1941 struct vcpu_svm *svm = to_svm(vcpu); local 1971 struct vcpu_svm *svm = to_svm(vcpu); local 1981 new_asid(struct vcpu_svm *svm, struct svm_cpu_data *sd) argument 1994 svm_set_dr6(struct vcpu_svm *svm, unsigned long value) argument 2009 struct vcpu_svm *svm = to_svm(vcpu); local 2030 struct vcpu_svm *svm = to_svm(vcpu); local 2041 struct vcpu_svm *svm = to_svm(vcpu); local 2054 struct vcpu_svm *svm = to_svm(vcpu); local 2069 struct vcpu_svm *svm = to_svm(vcpu); local 2101 struct vcpu_svm *svm = to_svm(vcpu); local 2189 struct vcpu_svm *svm = to_svm(vcpu); local 2214 struct vcpu_svm *svm = to_svm(vcpu); local 2255 struct vcpu_svm *svm = to_svm(vcpu); local 2346 struct vcpu_svm *svm = to_svm(vcpu); local 2369 struct vcpu_svm *svm = to_svm(vcpu); local 2407 svm_set_gif(struct vcpu_svm *svm, bool value) argument 2491 struct vcpu_svm *svm = to_svm(vcpu); local 2555 svm_clr_iret_intercept(struct vcpu_svm *svm) argument 2561 svm_set_iret_intercept(struct vcpu_svm *svm) argument 2569 struct vcpu_svm *svm = to_svm(vcpu); local 2605 struct vcpu_svm *svm = to_svm(vcpu); local 2628 struct vcpu_svm *svm = to_svm(vcpu); local 2702 struct vcpu_svm *svm = to_svm(vcpu); local 2737 struct vcpu_svm *svm = to_svm(vcpu); local 2826 struct vcpu_svm *svm = to_svm(vcpu); local 2934 struct vcpu_svm *svm = to_svm(vcpu); local 2948 struct vcpu_svm *svm = to_svm(vcpu); local 2973 struct vcpu_svm *svm = to_svm(vcpu); local 3216 struct vcpu_svm *svm = to_svm(vcpu); local 3312 struct vcpu_svm *svm = to_svm(vcpu); local 3490 struct vcpu_svm *svm = to_svm(vcpu); local 3534 struct vcpu_svm *svm = to_svm(vcpu); local 3557 struct vcpu_svm *svm = to_svm(vcpu); local 3578 struct vcpu_svm *svm = to_svm(vcpu); local 3588 struct vcpu_svm *svm = to_svm(vcpu); local 3611 struct vcpu_svm *svm = to_svm(vcpu); local 3683 struct vcpu_svm *svm = to_svm(vcpu); local 3706 struct vcpu_svm *svm = to_svm(vcpu); local 3716 struct vcpu_svm *svm = to_svm(vcpu); local 3735 struct vcpu_svm *svm = to_svm(vcpu); local 3752 struct vcpu_svm *svm = to_svm(vcpu); local 3767 struct vcpu_svm *svm = to_svm(vcpu); local 3793 struct vcpu_svm *svm = to_svm(vcpu); local 3813 struct vcpu_svm *svm = to_svm(vcpu); local 3843 struct vcpu_svm *svm = to_svm(vcpu); local 3887 struct vcpu_svm *svm = to_svm(vcpu); local 3942 struct vcpu_svm *svm = to_svm(vcpu); local 3949 struct vcpu_svm *svm = to_svm(vcpu); local 3962 struct vcpu_svm *svm = to_svm(vcpu); local 3979 struct vcpu_svm *svm = to_svm(vcpu); local 4008 struct vcpu_svm *svm = to_svm(vcpu); local 4077 struct vcpu_svm *svm = to_svm(vcpu); local 4106 struct vcpu_svm *svm = to_svm(vcpu); local 4124 struct vcpu_svm *svm = to_svm(vcpu); local 4253 struct vcpu_svm *svm = to_svm(vcpu); local 4312 struct vcpu_svm *svm = to_svm(vcpu); local 4436 struct vcpu_svm *svm = to_svm(vcpu); local 4572 struct vcpu_svm *svm = to_svm(vcpu); local 4583 struct vcpu_svm *svm = to_svm(vcpu); local 4599 struct vcpu_svm *svm = to_svm(vcpu); local 4651 struct vcpu_svm *svm = to_svm(vcpu); local 4713 struct vcpu_svm *svm = to_svm(vcpu); local 4874 struct vcpu_svm *svm = to_svm(vcpu); local [all...] |
H A D | svm.h | 22 #include <asm/svm.h> 314 void recalc_intercepts(struct vcpu_svm *svm); 403 static inline void set_exception_intercept(struct vcpu_svm *svm, u32 bit) argument 405 struct vmcb *vmcb = svm->vmcb01.ptr; 410 recalc_intercepts(svm); 413 static inline void clr_exception_intercept(struct vcpu_svm *svm, u32 bit) argument 415 struct vmcb *vmcb = svm->vmcb01.ptr; 420 recalc_intercepts(svm); 423 static inline void svm_set_intercept(struct vcpu_svm *svm, int bit) argument 425 struct vmcb *vmcb = svm 432 svm_clr_intercept(struct vcpu_svm *svm, int bit) argument 441 svm_is_intercept(struct vcpu_svm *svm, int bit) argument 446 nested_vgif_enabled(struct vcpu_svm *svm) argument 452 get_vgif_vmcb(struct vcpu_svm *svm) argument 463 enable_gif(struct vcpu_svm *svm) argument 473 disable_gif(struct vcpu_svm *svm) argument 483 gif_set(struct vcpu_svm *svm) argument 493 nested_npt_enabled(struct vcpu_svm *svm) argument 498 nested_vnmi_enabled(struct vcpu_svm *svm) argument 513 get_vnmi_vmcb_l1(struct vcpu_svm *svm) argument 524 is_vnmi_enabled(struct vcpu_svm *svm) argument 571 struct vcpu_svm *svm = to_svm(vcpu); local 576 nested_exit_on_smi(struct vcpu_svm *svm) argument 581 nested_exit_on_intr(struct vcpu_svm *svm) argument 586 nested_exit_on_nmi(struct vcpu_svm *svm) argument 602 nested_svm_simple_vmexit(struct vcpu_svm *svm, u32 exit_code) argument [all...] |
H A D | hyperv.h | 12 #include "svm.h" 17 struct vcpu_svm *svm = to_svm(vcpu); local 18 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments; 31 struct vcpu_svm *svm = to_svm(vcpu); local 32 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments;
|
H A D | avic.c | 28 #include "svm.h" 85 static void avic_activate_vmcb(struct vcpu_svm *svm) argument 87 struct vmcb *vmcb = svm->vmcb01.ptr; 101 if (x2avic_enabled && apic_x2apic_mode(svm->vcpu.arch.apic)) { 105 svm_set_x2apic_msr_interception(svm, false); 111 kvm_make_request(KVM_REQ_TLB_FLUSH_CURRENT, &svm->vcpu); 116 svm_set_x2apic_msr_interception(svm, true); 120 static void avic_deactivate_vmcb(struct vcpu_svm *svm) argument 122 struct vmcb *vmcb = svm->vmcb01.ptr; 131 if (is_guest_mode(&svm 244 avic_init_vmcb(struct vcpu_svm *svm, struct vmcb *vmcb) argument 281 struct vcpu_svm *svm = to_svm(vcpu); local 492 struct vcpu_svm *svm = to_svm(vcpu); local 597 struct vcpu_svm *svm = to_svm(vcpu); local 612 struct vcpu_svm *svm = to_svm(vcpu); local 631 struct vcpu_svm *svm = to_svm(vcpu); local 695 struct vcpu_svm *svm = to_svm(vcpu); local 719 avic_init_vcpu(struct vcpu_svm *svm) argument 749 struct vcpu_svm *svm = to_svm(vcpu); local 776 svm_ir_list_del(struct vcpu_svm *svm, struct amd_iommu_pi_data *pi) argument 792 svm_ir_list_add(struct vcpu_svm *svm, struct amd_iommu_pi_data *pi) argument 861 get_pi_vcpu_info(struct kvm *kvm, struct kvm_kernel_irq_routing_entry *e, struct vcpu_data *vcpu_info, struct vcpu_svm **svm) argument 920 struct vcpu_svm *svm = NULL; local 1006 struct vcpu_svm *svm = to_svm(vcpu); local 1032 struct vcpu_svm *svm = to_svm(vcpu); local 1075 struct vcpu_svm *svm = to_svm(vcpu); local 1114 struct vcpu_svm *svm = to_svm(vcpu); local [all...] |
H A D | sev.c | 30 #include "svm.h" 581 static int sev_es_sync_vmsa(struct vcpu_svm *svm) argument 583 struct sev_es_save_area *save = svm->sev_es.vmsa; 586 if (svm->vcpu.guest_debug || (svm->vmcb->save.dr7 & ~DR7_FIXED_1)) 595 memcpy(save, &svm->vmcb->save, sizeof(svm->vmcb->save)); 598 save->rax = svm->vcpu.arch.regs[VCPU_REGS_RAX]; 599 save->rbx = svm->vcpu.arch.regs[VCPU_REGS_RBX]; 600 save->rcx = svm 640 struct vcpu_svm *svm = to_svm(vcpu); local 2377 struct vcpu_svm *svm; local 2393 dump_ghcb(struct vcpu_svm *svm) argument 2418 sev_es_sync_to_ghcb(struct vcpu_svm *svm) argument 2437 sev_es_sync_from_ghcb(struct vcpu_svm *svm) argument 2491 sev_es_validate_vmgexit(struct vcpu_svm *svm) argument 2616 sev_es_unmap_ghcb(struct vcpu_svm *svm) argument 2648 pre_sev_run(struct vcpu_svm *svm, int cpu) argument 2672 setup_vmgexit_scratch(struct vcpu_svm *svm, bool sync, u64 len) argument 2756 set_ghcb_msr_bits(struct vcpu_svm *svm, u64 value, u64 mask, unsigned int pos) argument 2763 get_ghcb_msr_bits(struct vcpu_svm *svm, u64 mask, unsigned int pos) argument 2768 set_ghcb_msr(struct vcpu_svm *svm, u64 value) argument 2773 sev_handle_vmgexit_msr_protocol(struct vcpu_svm *svm) argument 2861 struct vcpu_svm *svm = to_svm(vcpu); local 2965 sev_es_string_io(struct vcpu_svm *svm, int size, unsigned int port, int in) argument 2986 sev_es_vcpu_after_set_cpuid(struct vcpu_svm *svm) argument 3017 sev_vcpu_after_set_cpuid(struct vcpu_svm *svm) argument 3031 sev_es_init_vmcb(struct vcpu_svm *svm) argument 3095 sev_init_vmcb(struct vcpu_svm *svm) argument 3110 sev_es_vcpu_reset(struct vcpu_svm *svm) argument 3163 struct vcpu_svm *svm = to_svm(vcpu); local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | svm.c | 3 * tools/testing/selftests/kvm/lib/x86_64/svm.c 5 * Largely inspired from KVM unit test svm.c 35 struct svm_test_data *svm = addr_gva2hva(vm, svm_gva); local 37 svm->vmcb = (void *)vm_vaddr_alloc_page(vm); 38 svm->vmcb_hva = addr_gva2hva(vm, (uintptr_t)svm->vmcb); 39 svm->vmcb_gpa = addr_gva2gpa(vm, (uintptr_t)svm->vmcb); 41 svm->save_area = (void *)vm_vaddr_alloc_page(vm); 42 svm 63 generic_svm_setup(struct svm_test_data *svm, void *guest_rip, void *guest_rsp) argument [all...] |
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm_perf.h | 5 #include <asm/svm.h>
|
/linux-master/tools/arch/x86/include/uapi/asm/ |
H A D | kvm_perf.h | 5 #include <asm/svm.h>
|
/linux-master/arch/powerpc/include/asm/ |
H A D | mem_encrypt.h | 11 #include <asm/svm.h>
|
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | svm_vmcall_test.c | 15 static void l2_guest_code(struct svm_test_data *svm) argument 20 static void l1_guest_code(struct svm_test_data *svm) argument 24 struct vmcb *vmcb = svm->vmcb; 27 generic_svm_setup(svm, l2_guest_code, 30 run_guest(vmcb, svm->vmcb_gpa);
|
H A D | svm_nested_shutdown_test.c | 15 static void l2_guest_code(struct svm_test_data *svm) argument 20 static void l1_guest_code(struct svm_test_data *svm, struct idt_entry *idt) argument 24 struct vmcb *vmcb = svm->vmcb; 26 generic_svm_setup(svm, l2_guest_code, 36 run_guest(vmcb, svm->vmcb_gpa);
|
H A D | svm_int_ctl_test.c | 33 static void l2_guest_code(struct svm_test_data *svm) argument 58 static void l1_guest_code(struct svm_test_data *svm) argument 62 struct vmcb *vmcb = svm->vmcb; 67 generic_svm_setup(svm, l2_guest_code, 80 run_guest(vmcb, svm->vmcb_gpa);
|
H A D | hyperv_svm_test.c | 69 static void __attribute__((__flatten__)) guest_code(struct svm_test_data *svm, argument 74 struct vmcb *vmcb = svm->vmcb; 83 GUEST_ASSERT(svm->vmcb_gpa); 85 generic_svm_setup(svm, l2_guest_code, 97 run_guest(vmcb, svm->vmcb_gpa); 104 __set_bit(2 * (MSR_FS_BASE & 0x1fff), svm->msr + 0x800); 105 run_guest(vmcb, svm->vmcb_gpa); 111 run_guest(vmcb, svm->vmcb_gpa); 116 __set_bit(2 * (MSR_GS_BASE & 0x1fff), svm->msr + 0x800); 119 run_guest(vmcb, svm [all...] |
H A D | nested_exceptions_test.c | 76 static void svm_run_l2(struct svm_test_data *svm, void *l2_code, int vector, argument 79 struct vmcb *vmcb = svm->vmcb; 83 run_guest(vmcb, svm->vmcb_gpa); 92 static void l1_svm_code(struct svm_test_data *svm) argument 94 struct vmcb_control_area *ctrl = &svm->vmcb->control; 97 generic_svm_setup(svm, NULL, &l2_guest_stack[L2_GUEST_STACK_SIZE]); 98 svm->vmcb->save.idtr.limit = 0; 102 svm_run_l2(svm, l2_ss_pending_test, SS_VECTOR, SS_ERROR_CODE); 103 svm_run_l2(svm, l2_ss_injected_gp_test, GP_VECTOR, GP_ERROR_CODE_AMD); 106 svm_run_l2(svm, l2_ss_injected_df_tes [all...] |
H A D | smm_test.c | 72 struct svm_test_data *svm = arg; local 87 generic_svm_setup(svm, l2_guest_code, 103 run_guest(svm->vmcb, svm->vmcb_gpa); 104 run_guest(svm->vmcb, svm->vmcb_gpa);
|
H A D | triple_fault_event_test.c | 43 void l1_guest_code_svm(struct svm_test_data *svm) argument 45 struct vmcb *vmcb = svm->vmcb; 47 generic_svm_setup(svm, l2_guest_code, 53 run_guest(vmcb, svm->vmcb_gpa);
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | cc_platform.c | 14 #include <asm/svm.h>
|
/linux-master/arch/x86/kvm/ |
H A D | Makefile | 24 kvm-amd-y += svm/svm.o svm/vmenter.o svm/pmu.o svm/nested.o svm/avic.o \ 25 svm/sev.o 26 kvm-amd-$(CONFIG_KVM_HYPERV) += svm/hyperv.o 31 kvm-amd-y += svm/svm_onhyperv.o 39 $(obj)/svm/vmente [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_svm.c | 78 #define SVM_DBG(s,f,a...) NV_DEBUG((s)->drm, "svm: "f"\n", ##a) 79 #define SVM_ERR(s,f,a...) NV_WARN((s)->drm, "svm: "f"\n", ##a) 94 nouveau_ivmm_find(struct nouveau_svm *svm, u64 inst) argument 97 list_for_each_entry(ivmm, &svm->inst, head) { 105 NV_DEBUG((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a) 107 NV_WARN((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a) 169 if (!cli->svm.svmm) { 186 nouveau_dmem_migrate_vma(cli->drm, cli->svm.svmm, vma, addr, 210 mutex_lock(&svmm->vmm->cli->drm->svm->mutex); 211 ivmm = nouveau_ivmm_find(svmm->vmm->cli->drm->svm, ins 381 nouveau_svm_fault_replay(struct nouveau_svm *svm) argument 396 nouveau_svm_fault_cancel(struct nouveau_svm *svm, u64 inst, u8 hub, u8 gpc, u8 client) argument 411 nouveau_svm_fault_cancel_fault(struct nouveau_svm *svm, struct nouveau_svm_fault *fault) argument 453 nouveau_svm_fault_cache(struct nouveau_svm *svm, struct nouveau_svm_fault_buffer *buffer, u32 offset) argument 716 struct nouveau_svm *svm = container_of(buffer, typeof(*svm), buffer[buffer->id]); local 939 nouveau_svm_fault_buffer_fini(struct nouveau_svm *svm, int id) argument 948 nouveau_svm_fault_buffer_init(struct nouveau_svm *svm, int id) argument 961 nouveau_svm_fault_buffer_dtor(struct nouveau_svm *svm, int id) argument 982 nouveau_svm_fault_buffer_ctor(struct nouveau_svm *svm, s32 oclass, int id) argument 1020 struct nouveau_svm *svm = drm->svm; local 1028 struct nouveau_svm *svm = drm->svm; local 1036 struct nouveau_svm *svm = drm->svm; local 1052 struct nouveau_svm *svm; local [all...] |
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | svm_util.h | 12 #include <asm/svm.h> 15 #include "svm.h" 60 void generic_svm_setup(struct svm_test_data *svm, void *guest_rip, void *guest_rsp);
|
/linux-master/drivers/iommu/intel/ |
H A D | svm.c | 47 svm_lookup_device_by_dev(struct intel_svm *svm, struct device *dev) argument 52 list_for_each_entry_rcu(t, &svm->devs, list) { 171 static void __flush_svm_range_dev(struct intel_svm *svm, argument 181 qi_flush_piotlb(sdev->iommu, sdev->did, svm->pasid, address, pages, ih); 184 svm->pasid, sdev->qdep, address, 187 svm->pasid, sdev->qdep); 191 static void intel_flush_svm_range_dev(struct intel_svm *svm, argument 202 __flush_svm_range_dev(svm, sdev, start, align >> VTD_PAGE_SHIFT, ih); 207 static void intel_flush_svm_range(struct intel_svm *svm, unsigned long address, argument 213 list_for_each_entry_rcu(sdev, &svm 218 intel_flush_svm_all(struct intel_svm *svm) argument 244 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); local 257 struct intel_svm *svm = container_of(mn, struct intel_svm, notifier); local 290 struct intel_svm *svm; local 324 struct intel_svm *svm; local 395 struct intel_svm *svm; local [all...] |
/linux-master/arch/arm/mm/ |
H A D | ioremap.c | 52 struct static_vm *svm; local 55 list_for_each_entry(svm, &static_vmlist, list) { 56 vm = &svm->vm; 66 return svm; 74 struct static_vm *svm; local 77 list_for_each_entry(svm, &static_vmlist, list) { 78 vm = &svm->vm; 85 return svm; 91 void __init add_static_vm_early(struct static_vm *svm) argument 97 vm = &svm 279 struct static_vm *svm; local 424 struct static_vm *svm; local [all...] |