/linux-master/arch/x86/power/ |
H A D | hibernate_asm_64.S | 33 movq mmu_cr4_features(%rip), %rax 34 movq %rax, %rdx 39 movq %rax, %cr4; # turn PGE back on 41 /* We don't restore %rax, it must be 0 anyway */ 42 movq $saved_context, %rax 43 movq pt_regs_sp(%rax), %rsp 44 movq pt_regs_bp(%rax), %rbp 45 movq pt_regs_si(%rax), %rsi 46 movq pt_regs_di(%rax), %rdi 47 movq pt_regs_bx(%rax), [all...] |
/linux-master/arch/x86/um/ |
H A D | stub_64.S | 12 mov 0x0(%rsp), %rax 16 mov %rax, 8(%rbx) 17 cmp $0, %rax 24 add %rax, %rsp 27 pop %rax 42 cmp %rcx, %rax 47 mov %rax, (%rbx)
|
/linux-master/arch/x86/kernel/acpi/ |
H A D | wakeup_64.S | 20 movq saved_magic(%rip), %rax 22 cmpq %rdx, %rax 43 movq saved_rip(%rip), %rax 45 jmp *%rax 54 movq $saved_context, %rax 55 movq %rsp, pt_regs_sp(%rax) 56 movq %rbp, pt_regs_bp(%rax) 57 movq %rsi, pt_regs_si(%rax) 58 movq %rdi, pt_regs_di(%rax) 59 movq %rbx, pt_regs_bx(%rax) [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | intel_pconfig.h | 52 unsigned long rax = MKTME_KEY_PROGRAM; local 58 : "=a" (rax), "=b" (key_program) 59 : "0" (rax), "1" (key_program) 62 return rax;
|
/linux-master/arch/x86/kernel/ |
H A D | ibt_selftest.S | 14 lea ibt_selftest_noendbr(%rip), %rax 16 jmp *%rax
|
H A D | relocate_kernel_64.S | 67 movq %cr0, %rax 68 movq %rax, CR0(%r11) 69 movq %cr3, %rax 70 movq %rax, CR3(%r11) 71 movq %cr4, %rax 72 movq %rax, CR4(%r11) 75 movq %rax, %r13 126 movq %cr4, %rax 127 andq $~(X86_CR4_CET), %rax 128 movq %rax, [all...] |
H A D | ftrace_64.S | 83 movq %rax, RAX(%rsp) 128 movq RAX(%rsp), %rax 183 movq RIP(%rsp), %rax 184 movq %rax, MCOUNT_REG_SIZE(%rsp) 247 movq EFLAGS(%rsp), %rax 248 movq %rax, MCOUNT_REG_SIZE(%rsp) 251 movq RIP(%rsp), %rax 252 movq %rax, MCOUNT_REG_SIZE+8(%rsp) 262 movq ORIG_RAX(%rsp), %rax 263 movq %rax, MCOUNT_REG_SIZ [all...] |
H A D | head_64.S | 82 leaq .Lon_kernel_cs(%rip), %rax 83 pushq %rax 115 addq %rcx, %rax 118 mov %rax, %rdi 133 movq %rax, %cr3 179 movq phys_base(%rip), %rax 180 addq $(init_top_pgt - __START_KERNEL_map), %rax 187 addq sme_me_mask(%rip), %rax 193 movq %rax, %cr3 302 movq trampoline_lock(%rip), %rax [all...] |
/linux-master/arch/x86/platform/efi/ |
H A D | efi_thunk_64.S | 37 movq %rsp, %rax 39 push %rax 45 movq 0x18(%rax), %rbp 46 movq 0x20(%rax), %rbx 47 movq 0x28(%rax), %rax 55 movq $__START_KERNEL_map, %rax 56 subq phys_base(%rip), %rax 60 subq %rax, %rbp 61 subq %rax, [all...] |
H A D | efi_stub_64.S | 17 mov 16(%rbp), %rax 20 mov %rax, 40(%rsp)
|
/linux-master/arch/x86/lib/ |
H A D | memset_64.S | 20 * rax original destination 38 movq %r9,%rax 51 movabs $0x0101010101010101,%rax 52 imulq %rcx,%rax 67 movq %rax,(%rdi) 68 movq %rax,8(%rdi) 69 movq %rax,16(%rdi) 70 movq %rax,24(%rdi) 71 movq %rax,32(%rdi) 72 movq %rax,4 [all...] |
H A D | clear_page_64.S | 31 #define PUT(x) movq %rax,x*8(%rdi) 32 movq %rax,(%rdi) 60 * rax is zero 86 1: movq %rax,(%rdi) 96 10: movq %rax,(%rdi) 97 11: movq %rax,8(%rdi) 98 12: movq %rax,16(%rdi) 99 13: movq %rax,24(%rdi) 100 14: movq %rax,32(%rdi) 101 15: movq %rax,4 [all...] |
H A D | hweight.S | 49 movabsq $0x5555555555555555, %rax 51 andq %rdx, %rax # t &= 0x5555555555555555 53 subq %rax, %rdi # w -= t 55 movq %rdi, %rax # w -> t 57 andq %rdx, %rax # t &= 0x3333333333333333 59 addq %rdx, %rax # w = w_tmp + t 61 movq %rax, %rdx # w -> t 63 addq %rdx, %rax # w_tmp += t 65 andq %rdx, %rax # w_tmp &= 0x0f0f0f0f0f0f0f0f 67 imulq %rdx, %rax # w_tm [all...] |
H A D | csum-copy_64.S | 64 /* r9: zero, r8: temp2, rbx: temp1, rax: sum, rcx: saved length */ 95 adcq %rbx, %rax 96 adcq %r8, %rax 97 adcq %r11, %rax 98 adcq %rdx, %rax 99 adcq %r10, %rax 100 adcq %r15, %rax 101 adcq %r14, %rax 102 adcq %r13, %rax 129 adcq %r9, %rax [all...] |
H A D | cmpxchg16b_emu.S | 13 * %rax : low 64 bits of old value 26 cmpq __percpu (%rsi), %rax 45 movq __percpu (%rsi), %rax
|
H A D | copy_user_64.S | 30 * it simpler for us, we can clobber rsi/rdi and rax freely. 57 2: movq (%rsi),%rax 58 3: movq %rax,(%rdi) 77 movq %rcx,%rax 86 1: leaq (%rax,%rcx,8),%rcx
|
/linux-master/tools/arch/x86/lib/ |
H A D | memset_64.S | 20 * rax original destination 38 movq %r9,%rax 51 movabs $0x0101010101010101,%rax 52 imulq %rcx,%rax 67 movq %rax,(%rdi) 68 movq %rax,8(%rdi) 69 movq %rax,16(%rdi) 70 movq %rax,24(%rdi) 71 movq %rax,32(%rdi) 72 movq %rax,4 [all...] |
/linux-master/arch/x86/entry/vsyscall/ |
H A D | vsyscall_emu_64.S | 20 mov $__NR_gettimeofday, %rax 26 mov $__NR_time, %rax 32 mov $__NR_getcpu, %rax
|
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | encls.h | 56 * @rax: function number 67 #define __encls_ret_N(rax, inputs...) \ 75 : "a"(rax), inputs \ 80 #define __encls_ret_1(rax, rcx) \ 82 __encls_ret_N(rax, "c"(rcx)); \ 85 #define __encls_ret_2(rax, rbx, rcx) \ 87 __encls_ret_N(rax, "b"(rbx), "c"(rcx)); \ 90 #define __encls_ret_3(rax, rbx, rcx, rdx) \ 92 __encls_ret_N(rax, "b"(rbx), "c"(rcx), "d"(rdx)); \ 97 * @rax [all...] |
/linux-master/arch/x86/entry/ |
H A D | entry_64_fred.S | 99 mov $__KERNEL_CS, %rax 100 push %rax 110 lea 1f(%rip), %rax 111 push %rax /* Return RIP */
|
H A D | entry_64_compat.S | 57 pushq %rax 58 SWITCH_TO_KERNEL_CR3 scratch_reg=%rax 59 popq %rax 87 pushq %rax /* pt_regs->orig_ax */ 88 PUSH_AND_CLEAR_REGS rax=$-ENOSYS 204 pushq %rax /* pt_regs->orig_ax */ 205 PUSH_AND_CLEAR_REGS rcx=%rbp rax=$-ENOSYS 235 popq %rax /* pt_regs->rax */
|
H A D | entry_64.S | 69 * rax system call number 107 pushq %rax /* pt_regs->orig_ax */ 109 PUSH_AND_CLEAR_REGS rax=$-ENOSYS 116 /* clobbers %rax, make sure it is after saving the syscall nr */ 222 * rax: prev task we switched from 240 movq %rax, %rdi /* prev */ 264 pushq %rax 270 popq %rax 298 ALTERNATIVE "call error_entry; movq %rax, %rsp", \ 484 movq %rax, [all...] |
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | xen_vmcall_test.c | 30 unsigned long rax = INPUTVALUE; local 41 "=a"(rax) : 42 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), 44 GUEST_ASSERT(rax == RETVALUE); 62 __asm__ __volatile__("call *%1" : "=a"(rax) : 64 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), 66 GUEST_ASSERT(rax == RETVALUE); 69 rax = 0; 72 __asm__ __volatile__("call *%1" : "=a"(rax) : 74 "a"(rax), " [all...] |
/linux-master/tools/testing/selftests/sgx/ |
H A D | test_encl_bootstrap.S | 47 lea __encl_base(%rip), %rax 48 sub %rax, %rbx 49 lea encl_stack(%rip), %rax 50 add %rbx, %rax 55 lea -1(%rbx), %rax 57 xchg %rsp, %rax 58 push %rax 72 pop %rax 73 mov %rax, %rsp 76 mov $4, %rax [all...] |
/linux-master/arch/x86/entry/vdso/ |
H A D | vsgx.S | 107 mov %rbx, %rax 118 push %rax 121 push %rax 127 * Load the callback pointer to %rax and lfence for LVI (load value 130 mov SGX_ENCLAVE_RUN_USER_HANDLER(%rax), %rax 132 call *%rax
|