/freebsd-11.0-release/crypto/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 233 #.set r10,10 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 279 # r9,r10, r11 are the equivalents of c1,c2, c3. 289 $UMULH r10,r5,r5 #in first iteration. No need 305 addc r10,r7,r10 # now add to temp result. 309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2; 315 addze r10,r0 323 addze r10,r10 [all...] |
H A D | x86_64-mont5.pl | 72 $lo0="%r10"; 112 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8) 114 and \$-1024,%r10 # minimize TLB usage 122 sub %r10,%r11 124 lea (%r10,%r11),%rsp 126 cmp %r10,%rsp 133 cmp %r10,%rsp 137 lea .Linc(%rip),%r10 147 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000 148 movdqa 16(%r10), [all...] |
H A D | rsaz-x86_64.pl | 170 movq %rdx, %r10 171 adcq \$0, %r10 174 addq %rax, %r10 219 addq %rax, %r10 262 movq %r10, %rdx 266 lea (%rcx,%r10,2), %r10 #shld \$1, %rcx, %r10 268 adcq %r11, %r11 #shld \$1, %r10, %r11 272 adcq %rdx, %r10 [all...] |
H A D | x86_64-mont.pl | 83 $lo0="%r10"; 128 lea -16(%rsp,$num,8),%r10 # future alloca(8*(num+2)) 130 and \$-1024,%r10 # minimize TLB usage 138 sub %r10,%r11 140 lea (%r10,%r11),%rsp 142 cmp %r10,%rsp 150 cmp %r10,%rsp 338 my @A=("%r10","%r11"); 363 lea -32(%rsp,$num,8),%r10 # future alloca(8*(num+4)) 365 and \$-1024,%r10 # minimiz [all...] |
/freebsd-11.0-release/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-ppc.pl | 176 ## Fills register %r10 -> .aes_consts (so you can -fPIC) 185 li r10, 0xd0 193 lvx $invhi, r12, r10 194 li r10, 0x110 201 lvx $sbot, r12, r10 202 li r10, 0x150 206 lvx $sb2t, r12, r10 222 ## Clobbers %xmm1-%xmm6, %r9, %r10, %r11, %rax 234 addi r10, r11, 0x40 247 lvx v1, r12, r11 # vmovdqa -0x40(%r11,%r10), [all...] |
H A D | vpaes-x86_64.pl | 79 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 101 lea .Lk_mc_backward(%rip),%r10 114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 116 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 157 movdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo 158 movdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16 162 movdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[] 189 lea .Lk_dsbd(%rip),%r10 196 add %r10, %r11 204 movdqa -0x20(%r10), [all...] |
/freebsd-11.0-release/secure/lib/libcrypto/amd64/ |
H A D | x86_64-mont5.S | 30 leaq -280(%rsp,%r9,8),%r10 32 andq $-1024,%r10 40 subq %r10,%r11 42 leaq (%r10,%r11,1),%rsp 44 cmpq %r10,%rsp 51 cmpq %r10,%rsp 55 leaq .Linc(%rip),%r10 60 movdqa 0(%r10),%xmm0 61 movdqa 16(%r10),%xmm1 62 leaq 24-112(%rsp,%r9,8),%r10 [all...] |
H A D | md5-x86_64.S | 45 leal -680876936(%rax,%r10,1),%eax 54 leal -389564586(%rdx,%r10,1),%edx 63 leal 606105819(%rcx,%r10,1),%ecx 72 leal -1044525330(%rbx,%r10,1),%ebx 81 leal -176418897(%rax,%r10,1),%eax 90 leal 1200080426(%rdx,%r10,1),%edx 99 leal -1473231341(%rcx,%r10,1),%ecx 108 leal -45705983(%rbx,%r10,1),%ebx 117 leal 1770035416(%rax,%r10,1),%eax 126 leal -1958414417(%rdx,%r10, [all...] |
H A D | x86_64-mont.S | 35 leaq -16(%rsp,%r9,8),%r10 37 andq $-1024,%r10 45 subq %r10,%r11 47 leaq (%r10,%r11,1),%rsp 49 cmpq %r10,%rsp 57 cmpq %r10,%rsp 73 movq %rax,%r10 76 imulq %r10,%rbp 80 addq %rax,%r10 94 movq %r10, [all...] |
H A D | rsaz-avx2.S | 289 movq 8(%rsp),%r10 312 addq %rax,%r10 315 addq %r9,%r10 320 movq %r10,%rax 336 addq %rax,%r10 347 shrq $29,%r10 351 addq %r10,%r11 458 movq 8(%rsp),%r10 471 addq %rax,%r10 479 leaq (%r9,%r10, [all...] |
H A D | rc4-x86_64.S | 20 xorq %r10,%r10 31 subq %r10,%rbx 33 movl (%rdi,%r10,4),%eax 39 leaq 1(%r10),%rsi 46 movl %edx,(%rdi,%r10,4) 50 movl (%rdi,%r10,4),%eax 57 leaq 1(%r10),%rsi 66 movl %edx,0(%rdi,%r10,4) 74 movl %edx,4(%rdi,%r10, [all...] |
H A D | ghash-x86_64.S | 33 movq %r9,%r10 37 shlq $60,%r10 43 xorq %r10,%r8 49 movq %r9,%r10 52 shlq $60,%r10 57 xorq %r10,%r8 64 movq %r9,%r10 67 shlq $60,%r10 72 xorq %r10,%r8 76 movq %r9,%r10 [all...] |
H A D | ecp_nistz256-x86_64.S | 35 movq 16(%rsi),%r10 40 adcq %r10,%r10 46 movq %r10,%rcx 48 sbbq 16(%rsi),%r10 56 cmovcq %rcx,%r10 59 movq %r10,16(%rdi) 78 movq 16(%rsi),%r10 86 movq %r10,%rcx 88 adcq 16(%rsi),%r10 [all...] |
H A D | rsaz-x86_64.S | 43 movq %rdx,%r10 44 adcq $0,%r10 47 addq %rax,%r10 92 addq %rax,%r10 135 movq %r10,%rdx 139 leaq (%rcx,%r10,2),%r10 145 adcq %rdx,%r10 149 movq %r10,24(%rsp) 178 movq %r12,%r10 [all...] |
H A D | sha512-x86_64.S | 47 movq 48(%rdi),%r10 65 xorq %r10,%r15 73 xorq %r10,%r15 113 addq %r10,%r12 126 movq %rax,%r10 132 xorq %r15,%r10 134 addq %r12,%r10 137 addq %r14,%r10 140 movq %r10,%r14 150 xorq %r10, [all...] |
H A D | vpaes-x86_64.S | 38 leaq .Lk_mc_backward(%rip),%r10 51 movdqa -64(%r11,%r10,1),%xmm1 53 movdqa (%r11,%r10,1),%xmm4 94 movdqa -96(%r10),%xmm4 95 movdqa -80(%r10),%xmm0 99 movdqa 64(%r11,%r10,1),%xmm1 126 leaq .Lk_dsbd(%rip),%r10 133 addq %r10,%r11 141 movdqa -32(%r10),%xmm4 142 movdqa -16(%r10), [all...] |
/freebsd-11.0-release/lib/libc/powerpc/sys/ |
H A D | setlogin.S | 37 mflr %r10 43 mtlr %r10
|
H A D | brk.S | 45 mflr %r10 48 mtlr %r10
|
H A D | sbrk.S | 44 mflr %r10 47 mtlr %r10
|
/freebsd-11.0-release/lib/libthr/arch/amd64/amd64/ |
H A D | _umtx_op_err.S | 35 #define KERNCALL movq %rcx, %r10; syscall
|
/freebsd-11.0-release/crypto/openssl/crypto/sha/asm/ |
H A D | sha512p8-ppc.pl | 131 li r10,`$FRAME+8*16+15` 133 stvx v20,r10,$sp # ABI says so 134 addi r10,r10,32 138 stvx v22,r10,$sp 139 addi r10,r10,32 142 stvx v24,r10,$sp 143 addi r10,r10,3 [all...] |
H A D | sha512-ia64.pl | 154 { .mib; add r10=2*$SZ,ctx 162 { .mmi; $LDW C_=[r10],4*$SZ 168 { .mmi; $LDW G_=[r10] 207 add r10=3*$SZ,input };; 212 $LDW X[12]=[r10],4*$SZ 218 $LDW X[ 8]=[r10],4*$SZ 224 $LDW X[ 4]=[r10],4*$SZ 230 $LDW X[ 0]=[r10],4*$SZ } 236 $LDW X[12]=[r10],4*$SZ 242 $LDW X[ 8]=[r10], [all...] |
/freebsd-11.0-release/contrib/gcclibs/libgomp/config/linux/x86/ |
H A D | futex.h | 41 register long r10 __asm__("%r10") = 0; 47 "d"(val), "r"(r10)
|
/freebsd-11.0-release/libexec/rtld-elf/amd64/ |
H A D | rtld_start.S | 75 * r10 0x8 109 pushq %r10 # Save %r10 111 .cfi_offset %r10,-88 128 popq %r10 # Restore %r10 130 .cfi_restore %r10
|
/freebsd-11.0-release/crypto/openssl/crypto/camellia/asm/ |
H A D | cmll-x86_64.pl | 478 shl \$32,%r10 # @S[2]|| 480 or %r11,%r10 # ||@S[3] 483 &_saveround (2,$out,-128,"%r8","%r10"); # KA<<<0 486 &_rotl128 ("%r8","%r10",15); 487 &_saveround (6,$out,-128,"%r8","%r10"); # KA<<<15 488 &_rotl128 ("%r8","%r10",15); # 15+15=30 489 &_saveround (8,$out,-128,"%r8","%r10"); # KA<<<30 492 &_rotl128 ("%r8","%r10",15); # 30+15=45 496 &_rotl128 ("%r8","%r10",15); # 45+15=60 497 &_saveround (14,$out,-128,"%r8","%r10"); # K [all...] |