/freebsd-11-stable/secure/lib/libcrypto/amd64/ |
H A D | x86_64-gf2m.S | 19 andq %rbp,%rax 23 andq %rbp,%rsi 27 andq %rbp,%rdi 68 andq %rbp,%rsi 70 shrq $4,%rbp 73 andq %rbp,%rdi 74 shrq $4,%rbp 78 andq %rbp,%rsi 79 shrq $4,%rbp 84 andq %rbp, [all...] |
H A D | wp-x86_64.S | 10 pushq %rbp 28 leaq .Ltable(%rip),%rbp 70 movq 4096(%rbp,%rsi,8),%r8 80 xorq 0(%rbp,%rsi,8),%r8 81 movq 7(%rbp,%rdi,8),%r9 87 movq 6(%rbp,%rsi,8),%r10 88 movq 5(%rbp,%rdi,8),%r11 94 movq 4(%rbp,%rsi,8),%r12 95 movq 3(%rbp,%rdi,8),%r13 101 movq 2(%rbp, [all...] |
H A D | cmll-x86_64.S | 23 pushq %rbp 34 leaq .LCamellia_SBOX(%rip),%rbp 60 movq 24(%rsp),%rbp 83 movl 2052(%rbp,%rsi,8),%edx 84 movl 0(%rbp,%rdi,8),%ecx 88 xorl 4(%rbp,%rsi,8),%edx 90 xorl 4(%rbp,%rdi,8),%ecx 93 xorl 0(%rbp,%rsi,8),%edx 94 xorl 2052(%rbp,%rdi,8),%ecx 97 xorl 2048(%rbp, [all...] |
H A D | x86_64-mont5.S | 22 pushq %rbp 196 movq %r8,%rbp 201 imulq %r10,%rbp 204 mulq %rbp 232 mulq %rbp 316 movq %r8,%rbp 324 imulq %r10,%rbp 327 mulq %rbp 358 mulq %rbp 420 movq -16(%rsi),%rbp [all...] |
H A D | x86_64-mont.S | 27 pushq %rbp 71 movq %r8,%rbp 76 imulq %r10,%rbp 79 mulq %rbp 107 mulq %rbp 132 movq %r8,%rbp 139 imulq %r10,%rbp 142 mulq %rbp 173 mulq %rbp 234 movq -16(%rsi),%rbp [all...] |
H A D | rsaz-x86_64.S | 12 pushq %rbp 35 movq %rax,%rbp 90 mulq %rbp 96 mulq %rbp 104 mulq %rbp 112 mulq %rbp 120 mulq %rbp 128 mulq %rbp 130 movq %rbp,%rax 144 movq 16(%rsi),%rbp [all...] |
H A D | aesni-mb-x86_64.S | 21 pushq %rbp 105 leaq 16(%rsp),%rbp 107 subq %rbx,%rbp 122 cmovgeq %rbp,%r8 123 cmovgq %rbp,%r12 130 cmovgeq %rbp,%r9 131 cmovgq %rbp,%r13 138 cmovgeq %rbp,%r10 139 cmovgq %rbp,%r14 146 cmovgeq %rbp, [all...] |
H A D | aes-x86_64.S | 23 movl 0(%r14,%rbp,8),%r12d 30 movl 0(%r14,%rbp,8),%r8d 37 xorl 3(%r14,%rbp,8),%r8d 48 xorl 2(%r14,%rbp,8),%r12d 55 xorl 2(%r14,%rbp,8),%r8d 62 xorl 1(%r14,%rbp,8),%r8d 77 movzbl 2(%r14,%rbp,8),%r12d 84 movl 0(%r14,%rbp,8),%ebp 111 movl 0(%r14,%rbp,8),%ebp 126 movl 2(%r14,%rbp, [all...] |
H A D | ecp_nistz256-x86_64.S | 357 pushq %rbp 392 popq %rbp 401 movq %rax,%rbp 405 movq %rbp,%rax 411 movq %rbp,%rax 417 movq %rbp,%rax 437 movq %r8,%rbp 440 shrq $32,%rbp 442 adcq %rbp,%r10 451 movq %rax,%rbp [all...] |
H A D | ghash-x86_64.S | 11 pushq %rbp 99 pushq %rbp 109 leaq 16+128(%rsp),%rbp 127 movq %r8,0(%rbp) 130 movq %rax,0-128(%rbp) 139 movq %r9,8(%rbp) 142 movq %rbx,8-128(%rbp) 151 movq %r8,16(%rbp) 154 movq %rax,16-128(%rbp) 163 movq %r9,24(%rbp) [all...] |
/freebsd-11-stable/sys/riscv/include/ |
H A D | machdep.h | 48 vm_offset_t fake_preload_metadata(struct riscv_bootparams *rbp);
|
/freebsd-11-stable/sys/amd64/include/ |
H A D | asm.h | 65 pushq %rbp; movq %rsp,%rbp; \ 67 popq %rbp; \ 70 pushq %rbp; movq %rsp,%rbp; \ 72 popq %rbp; \
|
/freebsd-11-stable/sys/cddl/dev/dtrace/amd64/ |
H A D | dtrace_asm.S | 49 movq TF_RBP(%rsp),%rbp; \ 90 * We must emulate a "pushq %rbp". To do this, we pull the stack 94 subq $16, %rsp /* make room for %rbp */ 103 subq $8, %rax /* make room for %rbp */ 108 movq %rbp, (%rax) /* store %rbp there */ 115 * We must emulate a "leave", which is the same as a "movq %rbp, %rsp" 116 * followed by a "popq %rbp". This is quite a bit simpler on amd64 125 movq (%rbp), %rax /* get new %rbp */ [all...] |
/freebsd-11-stable/sys/amd64/amd64/ |
H A D | locore.S | 71 movq %rsp, %rbp 72 movl 4(%rbp),%edi /* modulep (arg 1) */ 73 movl 8(%rbp),%esi /* kernend (arg 2) */
|
/freebsd-11-stable/crypto/openssl/crypto/md5/asm/ |
H A D | md5-x86_64.pl | 133 push %rbp 143 mov %rdi, %rbp # rbp = ctx 146 mov 0*4(%rbp), %eax # eax = ctx->A 147 mov 1*4(%rbp), %ebx # ebx = ctx->B 148 mov 2*4(%rbp), %ecx # ecx = ctx->C 149 mov 3*4(%rbp), %edx # edx = ctx->D 248 mov %eax, 0*4(%rbp) # ctx->A = A 249 mov %ebx, 1*4(%rbp) # ctx->B = B 250 mov %ecx, 2*4(%rbp) # ct [all...] |
/freebsd-11-stable/sys/amd64/vmm/amd/ |
H A D | svm_support.S | 37 * They are also responsible for saving/restoring the host %rbp across VMRUN. 39 #define VENTER push %rbp ; mov %rsp,%rbp 40 #define VLEAVE pop %rbp 63 * %rbp: because it is saved/restored by the function prologue/epilogue. 87 movq SCTX_RBP(%rsi), %rbp 111 movq %rbp, SCTX_RBP(%rax) 153 xor %rbp, %rbp
|
/freebsd-11-stable/crypto/openssl/crypto/bn/asm/ |
H A D | rsaz-x86_64.pl | 121 ($out, $inp, $mod) = ("%rdi", "%rsi", "%rbp"); # common internal API 135 push %rbp 162 mov %rax, %rbp # 8($inp) 217 mulq %rbp 223 mulq %rbp 231 mulq %rbp 239 mulq %rbp 247 mulq %rbp 255 mulq %rbp 257 movq %rbp, [all...] |
/freebsd-11-stable/crypto/openssl/crypto/aes/asm/ |
H A D | bsaes-x86_64.pl | 1161 push %rbp 1184 mov %rsp,%rbp # backup %rsp 1325 cmp %rax, %rbp 1328 lea (%rbp),%rsp # restore %rsp 1331 movaps 0x40(%rbp), %xmm6 1332 movaps 0x50(%rbp), %xmm7 1333 movaps 0x60(%rbp), %xmm8 1334 movaps 0x70(%rbp), %xmm9 1335 movaps 0x80(%rbp), %xmm10 1336 movaps 0x90(%rbp), [all...] |
/freebsd-11-stable/sys/dev/hatm/ |
H A D | if_hatm_intr.c | 308 he_intr_rbp(struct hatm_softc *sc, struct herbp *rbp, u_int large, argument 322 rbp->head = (READ4(sc, HE_REGO_RBP_S(large, group)) >> HE_REGS_RBP_HEAD) 323 & (rbp->size - 1); 326 if ((ntail = rbp->tail + 1) == rbp->size) 328 if (ntail == rbp->head) 348 m->m_data, rbp->bsize, hatm_mbuf_helper, 349 &rbp->rbp[rbp [all...] |
/freebsd-11-stable/sys/compat/ndis/ |
H A D | winx64_wrap.S | 94 push %rbp # insure that the stack 95 mov %rsp,%rbp # is 16-byte aligned 101 lea 56+8(%rbp),%rsi # source == old stack top (stack+56) 110 mov 40+8(%rbp),%r8 # set up arg4 (stack+40 -> %r8) 111 mov 48+8(%rbp),%r9 # set up arg5 (stack+48 -> %r9)
|
/freebsd-11-stable/sys/kern/ |
H A D | vfs_cluster.c | 96 struct buf *bp, *rbp, *reqbp; local 145 rbp = gbincore(&vp->v_bufobj, lblkno+i); 146 if (rbp == NULL || (rbp->b_flags & B_INVAL)) 156 && (0 == BUF_LOCK(rbp, 158 rbp->b_flags |= B_RAM; 159 BUF_UNLOCK(rbp); 273 rbp = cluster_rbuild(vp, filesize, lblkno, blkno, 275 lblkno += (rbp->b_bufsize / size); 276 if (rbp [all...] |
/freebsd-11-stable/sys/amd64/vmm/intel/ |
H A D | vmx_support.S | 44 #define VENTER push %rbp ; mov %rsp,%rbp 45 #define VLEAVE pop %rbp 59 movq %rbp,VMXCTX_GUEST_RBP(%rsp); \ 91 movq VMXCTX_GUEST_RBP(%rdi),%rbp; \ 124 movq %rbp, VMXCTX_HOST_RBP(%rdi); \ 133 movq VMXCTX_HOST_RBP(%rdi), %rbp; \
|
/freebsd-11-stable/lib/libc/amd64/gen/ |
H A D | _setjmp.S | 55 movq %rbp,24(%rax) /* 3; rbp */ 82 movq 24(%rdx),%rbp
|
H A D | sigsetjmp.S | 73 movq %rbp,24(%rcx) /* 3; rbp */ 103 movq 24(%rdx),%rbp
|
/freebsd-11-stable/stand/libsa/amd64/ |
H A D | _setjmp.S | 55 movq %rbp,24(%rax) /* 3; rbp */ 82 movq 24(%rdx),%rbp
|