/linux-master/tools/testing/selftests/kvm/ |
H A D | max_guest_memory_test.c | 78 sregs.cr0 ^= X86_CR0_WP;
|
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 913 static bool kvm_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument 916 if (cr0 & 0xffffffff00000000UL) 920 if ((cr0 & X86_CR0_NW) && !(cr0 & X86_CR0_CD)) 923 if ((cr0 & X86_CR0_PG) && !(cr0 & X86_CR0_PE)) 926 return static_call(kvm_x86_is_valid_cr0)(vcpu, cr0); 929 void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0) argument 939 if ((cr0 ^ old_cr0) == X86_CR0_WP) { 940 if (!(cr0 971 kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument [all...] |
H A D | smm.c | 58 CHECK_SMRAM32_OFFSET(cr0, 0xFFFC); 98 CHECK_SMRAM64_OFFSET(cr0, 0xFF58); 189 smram->cr0 = kvm_read_cr0(vcpu); 242 smram->cr0 = kvm_read_cr0(vcpu); 278 unsigned long cr0; local 318 cr0 = vcpu->arch.cr0 & ~(X86_CR0_PE | X86_CR0_EM | X86_CR0_TS | X86_CR0_PG); 319 static_call(kvm_x86_set_cr0)(vcpu, cr0); 415 u64 cr0, u64 cr3, u64 cr4) 440 bad = kvm_set_cr0(vcpu, cr0); 414 rsm_enter_protected_mode(struct kvm_vcpu *vcpu, u64 cr0, u64 cr3, u64 cr4) argument 567 unsigned long cr0; local [all...] |
H A D | emulate.c | 3458 ulong cr0; local 3460 cr0 = ctxt->ops->get_cr(ctxt, 0); 3461 cr0 &= ~X86_CR0_TS; 3462 ctxt->ops->set_cr(ctxt, 0, cr0);
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.h | 395 void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
|
H A D | vmx.c | 2481 vcpu->arch.cr0 &= ~guest_owned_bits; 2482 vcpu->arch.cr0 |= vmcs_readl(GUEST_CR0) & guest_owned_bits; 3258 static bool vmx_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument 3261 return nested_guest_cr0_valid(vcpu, cr0); 3264 return nested_host_cr0_valid(vcpu, cr0); 3269 void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument 3277 hw_cr0 = (cr0 & ~KVM_VM_CR0_ALWAYS_OFF); 3285 if (vmx->rmode.vm86_active && (cr0 & X86_CR0_PE)) 3288 if (!vmx->rmode.vm86_active && !(cr0 & X86_CR0_PE)) 3292 vmcs_writel(CR0_READ_SHADOW, cr0); 4293 unsigned long cr0, cr3, cr4; local [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 1848 static bool svm_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument 1853 void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) argument 1856 u64 hcr0 = cr0; 1861 if (!is_paging(vcpu) && (cr0 & X86_CR0_PG)) { 1867 if (is_paging(vcpu) && !(cr0 & X86_CR0_PG)) { 1874 vcpu->arch.cr0 = cr0; 1890 svm->vmcb->save.cr0 = hcr0; 1900 if (hcr0 == cr0) { 2606 unsigned long cr0 local 4455 unsigned long cr0, val; local [all...] |
H A D | svm.h | 120 u64 cr0; member in struct:vmcb_save_area_cached 549 void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 187 const unsigned long cr0; member in struct:kvm_mmu_role_regs 206 BUILD_MMU_ROLE_REGS_ACCESSOR(cr0, pg, X86_CR0_PG); 207 BUILD_MMU_ROLE_REGS_ACCESSOR(cr0, wp, X86_CR0_WP); 228 BUILD_MMU_ROLE_ACCESSOR(base, cr0, wp); 250 .cr0 = kvm_read_cr0_bits(vcpu, KVM_MMU_CR0_ROLE_BITS), 5132 /* Allow supervisor writes if !cr0.wp */ 5420 void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, unsigned long cr0, argument 5425 .cr0 = cr0,
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 751 unsigned long cr0; member in struct:kvm_vcpu_arch 1639 bool (*is_valid_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0); 1640 void (*set_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0); 2062 void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0); 2064 int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
|
H A D | suspend_32.h | 15 unsigned long cr0, cr2, cr3, cr4; member in struct:saved_context
|
H A D | hyperv-tlfs.h | 778 u64 cr0; member in struct:hv_init_vp_context
|
H A D | svm.h | 318 u64 cr0; member in struct:vmcb_save_area 375 u64 cr0; member in struct:sev_es_save_area
|
/linux-master/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 149 __u64 cr0, cr2, cr3, cr4, cr8; member in struct:kvm_sregs 160 __u64 cr0, cr2, cr3, cr4, cr8; member in struct:kvm_sregs2
|
/linux-master/arch/x86/kernel/ |
H A D | sev.c | 1055 vmsa->cr0 = AP_INIT_CR0_DEFAULT;
|
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | processor.h | 543 uint64_t cr0; local 545 __asm__ __volatile__("mov %%cr0, %[cr0]" 546 : /* output */ [cr0]"=r"(cr0)); 547 return cr0;
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 149 __u64 cr0, cr2, cr3, cr4, cr8; member in struct:kvm_sregs 160 __u64 cr0, cr2, cr3, cr4, cr8; member in struct:kvm_sregs2
|
/linux-master/arch/x86/boot/compressed/ |
H A D | efi_mixed.S | 192 movl %cr0, %eax 194 movl %eax, %cr0 235 movl %cr0, %eax 237 movl %eax, %cr0 284 movl %cr0, %eax 286 movl %eax, %cr0
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_vtl.c | 122 input->vp_context.cr0 = native_read_cr0();
|
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | processor.c | 105 sregs->cr0, sregs->cr2, sregs->cr3, sregs->cr4); 556 sregs.cr0 = X86_CR0_PE | X86_CR0_NE | X86_CR0_PG;
|
/linux-master/drivers/phy/freescale/ |
H A D | phy-fsl-lynx-28g.c | 32 #define LYNX_28G_PLLnCR0_REFCLK_SEL(cr0) (((cr0) & GENMASK(20, 16))) 113 u32 rstctl, cr0, cr1; member in struct:lynx_28g_pll 483 pll->cr0 = lynx_28g_pll_read(pll, PLLnCR0);
|
/linux-master/arch/powerpc/kexec/ |
H A D | relocate_32.S | 55 cmplwi cr0,r3,PVR_476FPE@h 57 cmplwi cr0,r3,PVR_476@h 59 cmplwi cr0,r3,PVR_476_ISS@h
|
/linux-master/arch/powerpc/include/asm/ |
H A D | ppc_asm.h | 588 #define cr0 0 macro
|
/linux-master/tools/testing/selftests/powerpc/primitives/asm/ |
H A D | ppc_asm.h | 588 #define cr0 0 macro
|
/linux-master/drivers/spi/ |
H A D | spi-rockchip.c | 521 u32 cr0 = CR0_FRF_SPI << CR0_FRF_OFFSET local 529 cr0 |= CR0_OPM_TARGET << CR0_OPM_OFFSET; 532 cr0 |= rs->rsd << CR0_RSD_OFFSET; 533 cr0 |= (spi->mode & 0x3U) << CR0_SCPH_OFFSET; 535 cr0 |= CR0_FBM_LSB << CR0_FBM_OFFSET; 537 cr0 |= BIT(spi_get_chipselect(spi, 0)) << CR0_SOI_OFFSET; 540 cr0 |= CR0_XFM_TR << CR0_XFM_OFFSET; 542 cr0 |= CR0_XFM_RO << CR0_XFM_OFFSET; 544 cr0 |= CR0_XFM_TO << CR0_XFM_OFFSET; 548 cr0 | 718 u32 cr0; local [all...] |