/linux-master/arch/x86/kernel/ |
H A D | sev_verify_cbit.S | 25 movq sme_me_mask(%rip), %rsi 26 testq %rsi, %rsi 30 movq sev_status(%rip), %rsi 31 testq %rsi, %rsi 34 /* Save CR4 in %rsi */ 35 movq %cr4, %rsi 38 movq %rsi, %rdx 71 movq %rsi, [all...] |
H A D | relocate_kernel_64.S | 50 * %rsi page_list 65 movq PTR(VA_CONTROL_PAGE)(%rsi), %r11 88 movq PTR(PA_CONTROL_PAGE)(%rsi), %r8 91 movq PTR(PA_TABLE_PAGE)(%rsi), %r9 94 movq PTR(PA_SWAP_PAGE)(%rsi), %r10 288 movq %rcx, %rsi /* For ever source page do a copy */ 289 andq $0xfffffffffffff000, %rsi 292 movq %rsi, %rax 299 movq %rdx, %rsi 304 movq %r10, %rsi [all...] |
/linux-master/arch/x86/lib/ |
H A D | cmpxchg16b_emu.S | 9 * Emulate 'cmpxchg16b %gs:(%rsi)' 12 * %rsi : memory location to compare 26 cmpq __percpu (%rsi), %rax 28 cmpq __percpu 8(%rsi), %rdx 32 movq %rbx, __percpu (%rsi) 33 movq %rcx, __percpu 8(%rsi) 45 movq __percpu (%rsi), %rax 46 movq __percpu 8(%rsi), %rdx
|
H A D | copy_page_64.S | 33 movq 0x8*0(%rsi), %rax 34 movq 0x8*1(%rsi), %rbx 35 movq 0x8*2(%rsi), %rdx 36 movq 0x8*3(%rsi), %r8 37 movq 0x8*4(%rsi), %r9 38 movq 0x8*5(%rsi), %r10 39 movq 0x8*6(%rsi), %r11 40 movq 0x8*7(%rsi), %r12 42 prefetcht0 5*64(%rsi) 53 leaq 64 (%rsi), [all...] |
H A D | memcpy_64.S | 18 * rsi source 65 movq 0*8(%rsi), %r8 66 movq 1*8(%rsi), %r9 67 movq 2*8(%rsi), %r10 68 movq 3*8(%rsi), %r11 69 leaq 4*8(%rsi), %rsi 84 addq %rdx, %rsi 94 movq -1*8(%rsi), %r8 95 movq -2*8(%rsi), [all...] |
H A D | memmove_64.S | 23 * rsi: src 34 cmp %rdi, %rsi 36 mov %rsi, %r8 67 movq 0*8(%rsi), %r11 68 movq 1*8(%rsi), %r10 69 movq 2*8(%rsi), %r9 70 movq 3*8(%rsi), %r8 71 leaq 4*8(%rsi), %rsi 87 movq -8(%rsi, [all...] |
H A D | csum-copy_64.S | 19 * rsi destination 107 movq %rbx, (%rsi) 109 movq %r8, 8(%rsi) 111 movq %r11, 16(%rsi) 113 movq %rdx, 24(%rsi) 116 movq %r10, 32(%rsi) 118 movq %r15, 40(%rsi) 120 movq %r14, 48(%rsi) 122 movq %r13, 56(%rsi) 125 leaq 64(%rsi), [all...] |
H A D | copy_user_uncached_64.S | 23 * rsi source 40 10: movq (%rsi),%r8 41 11: movq 8(%rsi),%r9 42 12: movq 16(%rsi),%r10 43 13: movq 24(%rsi),%r11 48 30: movq 32(%rsi),%r8 49 31: movq 40(%rsi),%r9 50 32: movq 48(%rsi),%r10 51 33: movq 56(%rsi),%r11 57 addq $64,%rsi [all...] |
H A D | copy_user_64.S | 21 * rsi source 30 * it simpler for us, we can clobber rsi/rdi and rax freely. 43 0: movb (%rsi),%al 46 inc %rsi 57 2: movq (%rsi),%rax 59 addq $8,%rsi
|
H A D | copy_mc_64.S | 37 movb (%rsi), %al 40 incq %rsi 52 movq (%rsi), %r8 55 addq $8, %rsi 68 movb (%rsi), %al 71 incq %rsi
|
/linux-master/arch/x86/virt/vmx/tdx/ |
H A D | tdxcall.S | 52 movq TDX_MODULE_rcx(%rsi), %rcx 53 movq TDX_MODULE_rdx(%rsi), %rdx 54 movq TDX_MODULE_r8(%rsi), %r8 55 movq TDX_MODULE_r9(%rsi), %r9 56 movq TDX_MODULE_r10(%rsi), %r10 57 movq TDX_MODULE_r11(%rsi), %r11 72 movq TDX_MODULE_r12(%rsi), %r12 73 movq TDX_MODULE_r13(%rsi), %r13 74 movq TDX_MODULE_r14(%rsi), %r14 75 movq TDX_MODULE_r15(%rsi), [all...] |
/linux-master/tools/arch/x86/lib/ |
H A D | memcpy_64.S | 17 * rsi source 64 movq 0*8(%rsi), %r8 65 movq 1*8(%rsi), %r9 66 movq 2*8(%rsi), %r10 67 movq 3*8(%rsi), %r11 68 leaq 4*8(%rsi), %rsi 83 addq %rdx, %rsi 93 movq -1*8(%rsi), %r8 94 movq -2*8(%rsi), [all...] |
/linux-master/arch/x86/um/ |
H A D | setjmp_64.S | 25 pop %rsi # Return address, and adjust the stack 29 push %rsi # Make the call/return stack happy 35 movq %rsi,56(%rdi) # Return address
|
H A D | stub_64.S | 31 pop %rsi
|
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_boot.S | 43 movq %rsi, %r11 /* Decrypted area */ 48 leaq __enc_copy(%rip), %rsi /* Encryption routine */ 54 movq %r11, %rsi /* Decrypted area */ 113 movq %rsi, %r11 /* Save decrypted area address */ 133 movq %r11, %rsi /* Source - decrypted area */ 138 movq %r8, %rsi /* Source - intermediate copy buffer */
|
/linux-master/arch/x86/boot/compressed/ |
H A D | idt_handlers_64.S | 24 pushq %rsi 42 movq ORIG_RAX(%rsp), %rsi 59 popq %rsi
|
/linux-master/arch/x86/crypto/ |
H A D | blake2s-core.S | 71 movd (%rsi,%rax,4),%xmm4 73 movd (%rsi,%rax,4),%xmm5 75 movd (%rsi,%rax,4),%xmm6 77 movd (%rsi,%rax,4),%xmm7 92 movd (%rsi,%rax,4),%xmm5 94 movd (%rsi,%rax,4),%xmm6 96 movd (%rsi,%rax,4),%xmm7 98 movd (%rsi,%rax,4),%xmm4 116 movd (%rsi,%rax,4),%xmm6 118 movd (%rsi, [all...] |
H A D | sm4-aesni-avx-asm_64.S | 151 * %rsi: dst (1..4 blocks) 235 vmovdqu RA0, 0*16(%rsi); 238 vmovdqu RA1, 1*16(%rsi); 240 vmovdqu RA2, 2*16(%rsi); 243 vmovdqu RA3, 3*16(%rsi); 367 * %rsi: dst (1..8 blocks) 396 vmovdqu RA0, (0 * 16)(%rsi); 397 vmovdqu RA1, (1 * 16)(%rsi); 398 vmovdqu RA2, (2 * 16)(%rsi); 399 vmovdqu RA3, (3 * 16)(%rsi); [all...] |
H A D | chacha-avx2-x86_64.S | 39 # %rsi: up to 2 data blocks output, o 145 vmovdqu %xmm6,0x00(%rsi) 152 vmovdqu %xmm6,0x10(%rsi) 159 vmovdqu %xmm6,0x20(%rsi) 166 vmovdqu %xmm6,0x30(%rsi) 174 vmovdqu %xmm6,0x40(%rsi) 180 vmovdqu %xmm6,0x50(%rsi) 186 vmovdqu %xmm6,0x60(%rsi) 192 vmovdqu %xmm6,0x70(%rsi) 205 mov %rsi, [all...] |
H A D | chacha-avx512vl-x86_64.S | 29 # %rsi: up to 2 data blocks output, o 118 vmovdqu %xmm6,0x00(%rsi) 125 vmovdqu %xmm6,0x10(%rsi) 132 vmovdqu %xmm6,0x20(%rsi) 139 vmovdqu %xmm6,0x30(%rsi) 147 vmovdqu %xmm6,0x40(%rsi) 153 vmovdqu %xmm6,0x50(%rsi) 159 vmovdqu %xmm6,0x60(%rsi) 165 vmovdqu %xmm6,0x70(%rsi) 186 vmovdqu8 %xmm1,(%rsi, [all...] |
/linux-master/arch/x86/include/uapi/asm/ |
H A D | ptrace.h | 63 unsigned long rsi; member in struct:pt_regs
|
/linux-master/arch/x86/include/asm/ |
H A D | GEN-for-each-reg.h | 12 GEN(rsi)
|
/linux-master/arch/x86/platform/efi/ |
H A D | efi_stub_64.S | 23 mov %rsi, %rcx
|
/linux-master/drivers/net/wireless/ |
H A D | Makefile | 19 obj-$(CONFIG_WLAN_VENDOR_RSI) += rsi/
|
/linux-master/arch/x86/purgatory/ |
H A D | entry64.S | 42 movq rsi(%rip), %rsi 68 rsi: .quad 0x0 label
|