/freebsd-11-stable/stand/powerpc/kboot/ |
H A D | kerneltramp.S | 55 mfmsr %r10 56 andi. %r10, %r10, 1 /* test MSR_LE */ 67 lwz %r10, 0(%r8) 68 mtctr %r10 75 li %r10, 4 76 lwbrx %r3, %r8, %r10 77 li %r10, 8 78 lwbrx %r4, %r8, %r10 79 li %r10, 1 [all...] |
/freebsd-11-stable/crypto/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 233 #.set r10,10 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 279 # r9,r10, r11 are the equivalents of c1,c2, c3. 289 $UMULH r10,r5,r5 #in first iteration. No need 305 addc r10,r7,r10 # now add to temp result. 309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2; 315 addze r10,r0 323 addze r10,r10 [all...] |
H A D | x86_64-mont5.pl | 72 $lo0="%r10"; 112 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8) 114 and \$-1024,%r10 # minimize TLB usage 122 sub %r10,%r11 124 lea (%r10,%r11),%rsp 126 cmp %r10,%rsp 133 cmp %r10,%rsp 137 lea .Linc(%rip),%r10 147 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000 148 movdqa 16(%r10), [all...] |
H A D | rsaz-x86_64.pl | 171 movq %rdx, %r10 172 adcq \$0, %r10 175 addq %rax, %r10 218 addq %rax, %r10 262 xorq %rbx, %rbx # rbx:r10:r9 = r10:r9 << 1 265 adcq %r10, %r10 274 adcq %rdx, %r10 278 movq %r10, 2 [all...] |
/freebsd-11-stable/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-ppc.pl | 176 ## Fills register %r10 -> .aes_consts (so you can -fPIC) 185 li r10, 0xd0 193 lvx $invhi, r12, r10 194 li r10, 0x110 201 lvx $sbot, r12, r10 202 li r10, 0x150 206 lvx $sb2t, r12, r10 222 ## Clobbers %xmm1-%xmm6, %r9, %r10, %r11, %rax 234 addi r10, r11, 0x40 247 lvx v1, r12, r11 # vmovdqa -0x40(%r11,%r10), [all...] |
/freebsd-11-stable/secure/lib/libcrypto/arm/ |
H A D | sha1-armv4-large.S | 34 ldrb r10,[r1,#2] 39 orr r9,r9,r10,lsl#8 40 eor r10,r5,r6 @ F_xx_xx 47 eor r10,r5,r6 @ F_xx_xx 53 and r10,r4,r10,ror#2 55 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 57 add r7,r7,r10 @ E+=F_00_19(B,C,D) 59 ldrb r10,[r [all...] |
H A D | aes-armv4.S | 174 sub r10,r3,#AES_encrypt-AES_Te @ Te 288 ldr r4,[r10,r7,lsl#2] @ Te3[s0>>0] 290 ldr r5,[r10,r8,lsl#2] @ Te2[s0>>8] 292 ldr r6,[r10,r9,lsl#2] @ Te1[s0>>16] 294 ldr r0,[r10,r0,lsl#2] @ Te0[s0>>24] 297 ldr r7,[r10,r7,lsl#2] @ Te1[s1>>16] 298 ldr r8,[r10,r8,lsl#2] @ Te3[s1>>0] 299 ldr r9,[r10,r9,lsl#2] @ Te2[s1>>8] 301 ldr r1,[r10,r1,lsl#2] @ Te0[s1>>24] 307 ldr r7,[r10,r [all...] |
/freebsd-11-stable/secure/lib/libcrypto/amd64/ |
H A D | x86_64-mont5.S | 30 leaq -280(%rsp,%r9,8),%r10 32 andq $-1024,%r10 40 subq %r10,%r11 42 leaq (%r10,%r11,1),%rsp 44 cmpq %r10,%rsp 51 cmpq %r10,%rsp 55 leaq .Linc(%rip),%r10 60 movdqa 0(%r10),%xmm0 61 movdqa 16(%r10),%xmm1 62 leaq 24-112(%rsp,%r9,8),%r10 [all...] |
H A D | md5-x86_64.S | 45 leal -680876936(%rax,%r10,1),%eax 54 leal -389564586(%rdx,%r10,1),%edx 63 leal 606105819(%rcx,%r10,1),%ecx 72 leal -1044525330(%rbx,%r10,1),%ebx 81 leal -176418897(%rax,%r10,1),%eax 90 leal 1200080426(%rdx,%r10,1),%edx 99 leal -1473231341(%rcx,%r10,1),%ecx 108 leal -45705983(%rbx,%r10,1),%ebx 117 leal 1770035416(%rax,%r10,1),%eax 126 leal -1958414417(%rdx,%r10, [all...] |
H A D | x86_64-mont.S | 35 leaq -16(%rsp,%r9,8),%r10 37 andq $-1024,%r10 45 subq %r10,%r11 47 leaq (%r10,%r11,1),%rsp 49 cmpq %r10,%rsp 57 cmpq %r10,%rsp 73 movq %rax,%r10 76 imulq %r10,%rbp 80 addq %rax,%r10 94 movq %r10, [all...] |
H A D | rsaz-avx2.S | 289 movq 8(%rsp),%r10 312 addq %rax,%r10 315 addq %r9,%r10 320 movq %r10,%rax 336 addq %rax,%r10 347 shrq $29,%r10 351 addq %r10,%r11 458 movq 8(%rsp),%r10 471 addq %rax,%r10 479 leaq (%r9,%r10, [all...] |
H A D | rc4-x86_64.S | 20 xorq %r10,%r10 31 subq %r10,%rbx 33 movl (%rdi,%r10,4),%eax 39 leaq 1(%r10),%rsi 46 movl %edx,(%rdi,%r10,4) 50 movl (%rdi,%r10,4),%eax 57 leaq 1(%r10),%rsi 66 movl %edx,0(%rdi,%r10,4) 74 movl %edx,4(%rdi,%r10, [all...] |
H A D | ghash-x86_64.S | 33 movq %r9,%r10 37 shlq $60,%r10 43 xorq %r10,%r8 49 movq %r9,%r10 52 shlq $60,%r10 57 xorq %r10,%r8 64 movq %r9,%r10 67 shlq $60,%r10 72 xorq %r10,%r8 76 movq %r9,%r10 [all...] |
H A D | ecp_nistz256-x86_64.S | 35 movq 16(%rsi),%r10 40 adcq %r10,%r10 46 movq %r10,%rcx 48 sbbq 16(%rsi),%r10 56 cmovcq %rcx,%r10 59 movq %r10,16(%rdi) 78 movq 16(%rsi),%r10 86 movq %r10,%rcx 88 adcq 16(%rsi),%r10 [all...] |
H A D | rsaz-x86_64.S | 44 movq %rdx,%r10 45 adcq $0,%r10 48 addq %rax,%r10 91 addq %rax,%r10 138 adcq %r10,%r10 147 adcq %rdx,%r10 151 movq %r10,24(%rsp) 200 movq 24(%rsi),%r10 211 mulq %r10 [all...] |
/freebsd-11-stable/lib/libc/powerpc/sys/ |
H A D | setlogin.S | 37 mflr %r10 43 mtlr %r10
|
/freebsd-11-stable/lib/libthr/arch/amd64/amd64/ |
H A D | _umtx_op_err.S | 35 #define KERNCALL movq %rcx, %r10; syscall
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/builtins/hexagon/ |
H A D | divdi3.S | 39 r10 = sub(r7,r6) // left shift count for bit & divisor define 44 r11 = add(r10,#1) // loop count is 1 more than shift count 45 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 46 r15:14 = lsl(r15:14,r10) // shift the bit left by same amount as divisor
|
H A D | moddi3.S | 37 r10 = sub(r7,r6) // left shift count for bit & divisor define 42 r11 = add(r10,#1) // loop count is 1 more than shift count 43 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 44 r15:14 = lsl(r15:14,r10) // shift the bit left by same amount as divisor
|
H A D | udivdi3.S | 31 r10 = sub(r7,r6) // left shift count for bit & divisor define 36 r11 = add(r10,#1) // loop count is 1 more than shift count 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 38 r15:14 = lsl(r15:14,r10) // shift the bit left by same amount as divisor
|
H A D | udivmoddi4.S | 31 r10 = sub(r7,r6) // left shift count for bit & divisor define 36 r11 = add(r10,#1) // loop count is 1 more than shift count 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 38 r15:14 = lsl(r15:14,r10) // shift the bit left by same amount as divisor
|
H A D | umoddi3.S | 31 r10 = sub(r7,r6) // left shift count for bit & divisor define 36 r11 = add(r10,#1) // loop count is 1 more than shift count 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 38 r15:14 = lsl(r15:14,r10) // shift the bit left by same amount as divisor
|
/freebsd-11-stable/crypto/openssl/crypto/sha/asm/ |
H A D | sha512p8-ppc.pl | 131 li r10,`$FRAME+8*16+15` 133 stvx v20,r10,$sp # ABI says so 134 addi r10,r10,32 138 stvx v22,r10,$sp 139 addi r10,r10,32 142 stvx v24,r10,$sp 143 addi r10,r10,3 [all...] |
/freebsd-11-stable/contrib/gcclibs/libgomp/config/linux/x86/ |
H A D | futex.h | 41 register long r10 __asm__("%r10") = 0; 47 "d"(val), "r"(r10)
|
/freebsd-11-stable/sys/contrib/cloudabi/ |
H A D | cloudabi_vdso_x86_64.S | 119 mov %rcx, %r10 132 mov %rcx, %r10 205 mov %rcx, %r10 218 mov %rcx, %r10 225 mov %rcx, %r10 232 mov %rcx, %r10 245 mov %rcx, %r10 258 mov %rcx, %r10 271 mov %rcx, %r10 290 mov %rcx, %r10 [all...] |