/linux-master/arch/x86/kernel/ |
H A D | alternative.c | 1643 ALTERNATIVE("", "lea %[mem], %%" _ASM_ARG1 "; call __alt_reloc_selftest;", X86_FEATURE_ALWAYS)
|
/linux-master/arch/arm64/include/asm/ |
H A D | io.h | 56 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", 68 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", 79 asm volatile(ALTERNATIVE("ldr %w0, [%1]", 90 asm volatile(ALTERNATIVE("ldr %0, [%1]",
|
H A D | kvm_hyp.h | 45 asm volatile(ALTERNATIVE(__mrs_s("%0", r##nvh), \ 55 asm volatile(ALTERNATIVE(__msr_s(r##nvh, "%x0"), \
|
H A D | tlbflush.h | 34 ALTERNATIVE("nop\n nop", \ 42 ALTERNATIVE("nop\n nop", \
|
H A D | sysreg.h | 1178 asm(ALTERNATIVE("nop", "dmb sy", ARM64_WORKAROUND_1508412)); \ 1180 asm(ALTERNATIVE("nop", "dmb sy", ARM64_WORKAROUND_1508412)); \
|
/linux-master/drivers/perf/ |
H A D | riscv_pmu_sbi.c | 41 asm volatile(ALTERNATIVE( \
|
/linux-master/arch/x86/kernel/fpu/ |
H A D | xstate.h | 111 * The 661 label is defined in the ALTERNATIVE* macros as the address of the 133 asm volatile(ALTERNATIVE(XRSTOR, \
|
/linux-master/arch/arm64/kvm/hyp/vhe/ |
H A D | switch.c | 161 asm(ALTERNATIVE("nop", "isb", ARM64_WORKAROUND_SPECULATIVE_AT));
|
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | tlb.c | 115 asm(ALTERNATIVE("isb", "nop", ARM64_WORKAROUND_SPECULATIVE_AT));
|
H A D | mem_protect.c | 316 asm(ALTERNATIVE("isb", "nop", ARM64_WORKAROUND_SPECULATIVE_AT));
|
/linux-master/arch/arm64/kvm/hyp/include/hyp/ |
H A D | switch.h | 686 asm(ALTERNATIVE("nop", "dmb sy", ARM64_WORKAROUND_1508412));
|
/linux-master/tools/arch/arm64/include/asm/ |
H A D | sysreg.h | 824 asm(ALTERNATIVE("nop", "dmb sy", ARM64_WORKAROUND_1508412)); \ 826 asm(ALTERNATIVE("nop", "dmb sy", ARM64_WORKAROUND_1508412)); \
|
/linux-master/arch/x86/kernel/cpu/ |
H A D | amd.c | 1230 asm volatile(ALTERNATIVE("", "div %2\n\t", X86_BUG_DIV0)
|
/linux-master/arch/x86/include/asm/ |
H A D | percpu.h | 292 asm qual (ALTERNATIVE("call this_cpu_cmpxchg8b_emu", \ 322 asm qual (ALTERNATIVE("call this_cpu_cmpxchg8b_emu", \ 361 asm qual (ALTERNATIVE("call this_cpu_cmpxchg16b_emu", \ 391 asm qual (ALTERNATIVE("call this_cpu_cmpxchg16b_emu", \
|
H A D | cmpxchg_32.h | 94 asm volatile(ALTERNATIVE(_lock_loc \ 123 asm volatile(ALTERNATIVE(_lock_loc \
|
H A D | alternative.h | 217 #define ALTERNATIVE(oldinstr, newinstr, ft_flags) \ macro 269 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory") 286 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 291 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 296 asm_inline volatile (ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \ 391 .macro ALTERNATIVE oldinstr, newinstr, ft_flags 425 * Same as ALTERNATIVE macro above but for two alternatives. If CPU
|
H A D | barrier.h | 15 #define mb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "mfence", \ 17 #define rmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "lfence", \ 19 #define wmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "sfence", \
|
H A D | nospec-branch.h | 313 ALTERNATIVE "", \ 326 ALTERNATIVE "", __stringify(verw _ASM_RIP(mds_verw_sel)), X86_FEATURE_CLEAR_CPU_BUF 331 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_LOOP 335 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_LOOP_ON_VMEXIT 396 ALTERNATIVE("", \ 515 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature])
|
/linux-master/arch/x86/entry/ |
H A D | entry_64_compat.S | 216 ALTERNATIVE "testb %al, %al; jz swapgs_restore_regs_and_return_to_usermode", \
|
H A D | entry_64.S | 130 ALTERNATIVE "testb %al, %al; jz swapgs_restore_regs_and_return_to_usermode", \ 254 ALTERNATIVE "jmp swapgs_restore_regs_and_return_to_usermode", \ 298 ALTERNATIVE "call error_entry; movq %rax, %rsp", \ 563 ALTERNATIVE "", "jmp xenpv_restore_regs_and_return_to_usermode", X86_FEATURE_XENPV 566 ALTERNATIVE "", "jmp .Lpti_restore_regs_and_return_to_usermode", X86_FEATURE_PTI 751 2: ALTERNATIVE "", "mfence", X86_BUG_SWAPGS_FENCE 764 ALTERNATIVE "", "ZAP_GS", X86_BUG_NULL_SEG 898 ALTERNATIVE "jmp .Lparanoid_entry_checkgs", "", X86_FEATURE_FSGSBASE 984 ALTERNATIVE "jmp .Lparanoid_exit_checkgs", "", X86_FEATURE_FSGSBASE 1420 ALTERNATIVE "jm [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | errata_list.h | 34 ALTERNATIVE(__stringify(RISCV_PTR do_trap_insn_fault), \ 40 ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \ 47 asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 76 asm volatile(ALTERNATIVE( \ 97 asm volatile(ALTERNATIVE( \
|
/linux-master/arch/x86/lib/ |
H A D | retpoline.S | 329 ALTERNATIVE JMP_RETBLEED_UNTRAIN_RET, JMP_SRSO_UNTRAIN_RET, X86_FEATURE_SRSO 377 * The ALTERNATIVE below adds a really loud warning to catch the case 388 ALTERNATIVE __stringify(ANNOTATE_UNRET_SAFE; ret), \
|
/linux-master/arch/s390/kernel/ |
H A D | entry.S | 35 ALTERNATIVE "nop", ".insn s,0xb2010000,\address", 193 39 ALTERNATIVE "nop", ".insn s,0xb2000000,\address", 193 43 ALTERNATIVE "b \lpswe; nopr", ".insn siy,0xeb0000000071,\address,0", 193 47 ALTERNATIVE "brcl 0,0", __stringify(mvc __PT_LAST_BREAK(8,\reg),__LC_LAST_BREAK), 193 103 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,12,0", 82 107 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,13,0", 82 111 ALTERNATIVE "TSTMSK \tif_ptr,\tif_mask; jz .+8; .insn rrf,0xb2e80000,0,0,13,0", \ 117 ALTERNATIVE "jz .+8; .insn rrf,0xb2e80000,0,0,12,0", \ 172 ALTERNATIVE "nop", "lpp _LPP_OFFSET", 40 542 0: ALTERNATIVE "no [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmenter.S | 111 ALTERNATIVE "jmp .Lspec_ctrl_done", "", X86_FEATURE_MSR_SPEC_CTRL
|
/linux-master/arch/riscv/lib/ |
H A D | csum.c | 56 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0, 173 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0, 260 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0,
|