Searched refs:r10 (Results 1 - 25 of 148) sorted by relevance

123456

/freebsd-11.0-release/crypto/openssl/crypto/bn/asm/
H A Dppc.pl233 #.set r10,10
275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows:
279 # r9,r10, r11 are the equivalents of c1,c2, c3.
289 $UMULH r10,r5,r5 #in first iteration. No need
305 addc r10,r7,r10 # now add to temp result.
309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2;
315 addze r10,r0
323 addze r10,r10
[all...]
H A Dx86_64-mont5.pl72 $lo0="%r10";
112 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8)
114 and \$-1024,%r10 # minimize TLB usage
122 sub %r10,%r11
124 lea (%r10,%r11),%rsp
126 cmp %r10,%rsp
133 cmp %r10,%rsp
137 lea .Linc(%rip),%r10
147 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000
148 movdqa 16(%r10),
[all...]
H A Drsaz-x86_64.pl170 movq %rdx, %r10
171 adcq \$0, %r10
174 addq %rax, %r10
219 addq %rax, %r10
262 movq %r10, %rdx
266 lea (%rcx,%r10,2), %r10 #shld \$1, %rcx, %r10
268 adcq %r11, %r11 #shld \$1, %r10, %r11
272 adcq %rdx, %r10
[all...]
H A Dx86_64-mont.pl83 $lo0="%r10";
128 lea -16(%rsp,$num,8),%r10 # future alloca(8*(num+2))
130 and \$-1024,%r10 # minimize TLB usage
138 sub %r10,%r11
140 lea (%r10,%r11),%rsp
142 cmp %r10,%rsp
150 cmp %r10,%rsp
338 my @A=("%r10","%r11");
363 lea -32(%rsp,$num,8),%r10 # future alloca(8*(num+4))
365 and \$-1024,%r10 # minimiz
[all...]
/freebsd-11.0-release/crypto/openssl/crypto/aes/asm/
H A Dvpaes-ppc.pl176 ## Fills register %r10 -> .aes_consts (so you can -fPIC)
185 li r10, 0xd0
193 lvx $invhi, r12, r10
194 li r10, 0x110
201 lvx $sbot, r12, r10
202 li r10, 0x150
206 lvx $sb2t, r12, r10
222 ## Clobbers %xmm1-%xmm6, %r9, %r10, %r11, %rax
234 addi r10, r11, 0x40
247 lvx v1, r12, r11 # vmovdqa -0x40(%r11,%r10),
[all...]
H A Dvpaes-x86_64.pl79 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax
101 lea .Lk_mc_backward(%rip),%r10
114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[]
116 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[]
157 movdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo
158 movdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16
162 movdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[]
189 lea .Lk_dsbd(%rip),%r10
196 add %r10, %r11
204 movdqa -0x20(%r10),
[all...]
/freebsd-11.0-release/secure/lib/libcrypto/amd64/
H A Dx86_64-mont5.S30 leaq -280(%rsp,%r9,8),%r10
32 andq $-1024,%r10
40 subq %r10,%r11
42 leaq (%r10,%r11,1),%rsp
44 cmpq %r10,%rsp
51 cmpq %r10,%rsp
55 leaq .Linc(%rip),%r10
60 movdqa 0(%r10),%xmm0
61 movdqa 16(%r10),%xmm1
62 leaq 24-112(%rsp,%r9,8),%r10
[all...]
H A Dmd5-x86_64.S45 leal -680876936(%rax,%r10,1),%eax
54 leal -389564586(%rdx,%r10,1),%edx
63 leal 606105819(%rcx,%r10,1),%ecx
72 leal -1044525330(%rbx,%r10,1),%ebx
81 leal -176418897(%rax,%r10,1),%eax
90 leal 1200080426(%rdx,%r10,1),%edx
99 leal -1473231341(%rcx,%r10,1),%ecx
108 leal -45705983(%rbx,%r10,1),%ebx
117 leal 1770035416(%rax,%r10,1),%eax
126 leal -1958414417(%rdx,%r10,
[all...]
H A Dx86_64-mont.S35 leaq -16(%rsp,%r9,8),%r10
37 andq $-1024,%r10
45 subq %r10,%r11
47 leaq (%r10,%r11,1),%rsp
49 cmpq %r10,%rsp
57 cmpq %r10,%rsp
73 movq %rax,%r10
76 imulq %r10,%rbp
80 addq %rax,%r10
94 movq %r10,
[all...]
H A Drsaz-avx2.S289 movq 8(%rsp),%r10
312 addq %rax,%r10
315 addq %r9,%r10
320 movq %r10,%rax
336 addq %rax,%r10
347 shrq $29,%r10
351 addq %r10,%r11
458 movq 8(%rsp),%r10
471 addq %rax,%r10
479 leaq (%r9,%r10,
[all...]
H A Drc4-x86_64.S20 xorq %r10,%r10
31 subq %r10,%rbx
33 movl (%rdi,%r10,4),%eax
39 leaq 1(%r10),%rsi
46 movl %edx,(%rdi,%r10,4)
50 movl (%rdi,%r10,4),%eax
57 leaq 1(%r10),%rsi
66 movl %edx,0(%rdi,%r10,4)
74 movl %edx,4(%rdi,%r10,
[all...]
H A Dghash-x86_64.S33 movq %r9,%r10
37 shlq $60,%r10
43 xorq %r10,%r8
49 movq %r9,%r10
52 shlq $60,%r10
57 xorq %r10,%r8
64 movq %r9,%r10
67 shlq $60,%r10
72 xorq %r10,%r8
76 movq %r9,%r10
[all...]
H A Decp_nistz256-x86_64.S35 movq 16(%rsi),%r10
40 adcq %r10,%r10
46 movq %r10,%rcx
48 sbbq 16(%rsi),%r10
56 cmovcq %rcx,%r10
59 movq %r10,16(%rdi)
78 movq 16(%rsi),%r10
86 movq %r10,%rcx
88 adcq 16(%rsi),%r10
[all...]
H A Drsaz-x86_64.S43 movq %rdx,%r10
44 adcq $0,%r10
47 addq %rax,%r10
92 addq %rax,%r10
135 movq %r10,%rdx
139 leaq (%rcx,%r10,2),%r10
145 adcq %rdx,%r10
149 movq %r10,24(%rsp)
178 movq %r12,%r10
[all...]
H A Dsha512-x86_64.S47 movq 48(%rdi),%r10
65 xorq %r10,%r15
73 xorq %r10,%r15
113 addq %r10,%r12
126 movq %rax,%r10
132 xorq %r15,%r10
134 addq %r12,%r10
137 addq %r14,%r10
140 movq %r10,%r14
150 xorq %r10,
[all...]
H A Dvpaes-x86_64.S38 leaq .Lk_mc_backward(%rip),%r10
51 movdqa -64(%r11,%r10,1),%xmm1
53 movdqa (%r11,%r10,1),%xmm4
94 movdqa -96(%r10),%xmm4
95 movdqa -80(%r10),%xmm0
99 movdqa 64(%r11,%r10,1),%xmm1
126 leaq .Lk_dsbd(%rip),%r10
133 addq %r10,%r11
141 movdqa -32(%r10),%xmm4
142 movdqa -16(%r10),
[all...]
/freebsd-11.0-release/lib/libc/powerpc/sys/
H A Dsetlogin.S37 mflr %r10
43 mtlr %r10
H A Dbrk.S45 mflr %r10
48 mtlr %r10
H A Dsbrk.S44 mflr %r10
47 mtlr %r10
/freebsd-11.0-release/lib/libthr/arch/amd64/amd64/
H A D_umtx_op_err.S35 #define KERNCALL movq %rcx, %r10; syscall
/freebsd-11.0-release/crypto/openssl/crypto/sha/asm/
H A Dsha512p8-ppc.pl131 li r10,`$FRAME+8*16+15`
133 stvx v20,r10,$sp # ABI says so
134 addi r10,r10,32
138 stvx v22,r10,$sp
139 addi r10,r10,32
142 stvx v24,r10,$sp
143 addi r10,r10,3
[all...]
H A Dsha512-ia64.pl154 { .mib; add r10=2*$SZ,ctx
162 { .mmi; $LDW C_=[r10],4*$SZ
168 { .mmi; $LDW G_=[r10]
207 add r10=3*$SZ,input };;
212 $LDW X[12]=[r10],4*$SZ
218 $LDW X[ 8]=[r10],4*$SZ
224 $LDW X[ 4]=[r10],4*$SZ
230 $LDW X[ 0]=[r10],4*$SZ }
236 $LDW X[12]=[r10],4*$SZ
242 $LDW X[ 8]=[r10],
[all...]
/freebsd-11.0-release/contrib/gcclibs/libgomp/config/linux/x86/
H A Dfutex.h41 register long r10 __asm__("%r10") = 0;
47 "d"(val), "r"(r10)
/freebsd-11.0-release/libexec/rtld-elf/amd64/
H A Drtld_start.S75 * r10 0x8
109 pushq %r10 # Save %r10
111 .cfi_offset %r10,-88
128 popq %r10 # Restore %r10
130 .cfi_restore %r10
/freebsd-11.0-release/crypto/openssl/crypto/camellia/asm/
H A Dcmll-x86_64.pl478 shl \$32,%r10 # @S[2]||
480 or %r11,%r10 # ||@S[3]
483 &_saveround (2,$out,-128,"%r8","%r10"); # KA<<<0
486 &_rotl128 ("%r8","%r10",15);
487 &_saveround (6,$out,-128,"%r8","%r10"); # KA<<<15
488 &_rotl128 ("%r8","%r10",15); # 15+15=30
489 &_saveround (8,$out,-128,"%r8","%r10"); # KA<<<30
492 &_rotl128 ("%r8","%r10",15); # 30+15=45
496 &_rotl128 ("%r8","%r10",15); # 45+15=60
497 &_saveround (14,$out,-128,"%r8","%r10"); # K
[all...]

Completed in 107 milliseconds

123456