/linux-master/arch/arm/lib/ |
H A D | csumpartialcopygeneric.S | 43 .Ldst_16bit: load2b r8, ip 45 adcs sum, sum, r8, put_byte_0 46 strb r8, [dst], #1 70 1: load2b r8, ip 72 adcs sum, sum, r8, put_byte_0 73 strb r8, [dst], #1 82 load1b r8 83 adcs sum, sum, r8, put_byte_0 @ update checksum 84 strb r8, [dst], #1 176 1: load4l r5, r6, r7, r8 [all...] |
H A D | copy_template.S | 86 stmfd sp!, {r5, r6, r8, r9} 105 4: ldr8w r1, r3, r4, r5, r6, r8, r9, ip, lr, abort=20f 107 str8w r0, r3, r4, r5, r6, r8, r9, ip, lr, abort=20f 127 ldr1w r1, r8, abort=20f 145 str1w r0, r8, abort=20f 151 7: ldmfd sp!, {r5, r6, r8, r9} 194 11: stmfd sp!, {r5, r6, r8 - r10} 204 13: ldr4w r1, r4, r5, r6, r8, abort=19f 214 orr r6, r6, r8, lspush #\push 215 mov r8, r [all...] |
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_boot.S | 55 movq %r8, %rdx /* Pagetables used for encryption */ 57 movq %rax, %r8 /* Workarea encryption routine */ 58 addq $PAGE_SIZE, %r8 /* Workarea intermediate copy buffer */ 134 movq %r8, %rdi /* Dest - intermediate copy buffer */ 138 movq %r8, %rsi /* Source - intermediate copy buffer */
|
/linux-master/arch/s390/kernel/ |
H A D | relocate_kernel.S | 22 * %r8 holds the source address 54 lgr %r8,%r5 # r8 = r5 55 nill %r8,0xf000 # masking 56 0: mvcle %r6,%r8,0x0 # copy PAGE_SIZE bytes from r8 to r6 - pad with 0
|
/linux-master/arch/powerpc/kernel/ |
H A D | 85xx_entry_mapping.S | 84 li r8,-1 86 slw r6,r8,r6 /* convert to mask */ 91 mfspr r8,SPRN_MAS3 95 and r8,r6,r8 99 or r25,r8,r9 100 ori r8,r25,(MAS3_SX|MAS3_SW|MAS3_SR) 115 mtspr SPRN_MAS3,r8 169 mtspr SPRN_MAS3,r8 223 mfspr r8,SPRN_MAS [all...] |
H A D | tm.S | 230 SAVE_GPRS(8, 10, r7) /* user r8-r10 */ 235 ld r8, GPR13(r1) /* user r13 */ 240 std r8, GPR13(r7) 319 ld r8, TM_FRAME_L1(r1) 320 mtspr SPRN_AMR, r8 371 mfspr r8, SPRN_AMR 372 std r8, TM_FRAME_L0(r1) 405 addi r8, r3, THREAD_CKVRSTATE 407 lvx v0, r8, r5 409 REST_32VRS(0, r5, r8) /* r [all...] |
/linux-master/arch/arm/mach-imx/ |
H A D | suspend-imx6.S | 103 ldr r8, [r7], #0x4 105 str r9, [r11, r8] 164 add r8, r1, r4 165 add r9, r8, r7 210 ldr r8, =PM_INFO_MMDC_IO_VAL_OFFSET 211 add r8, r8, r0 216 ldr r9, [r8], #0x8 224 ldr r9, [r8], #0x8 226 ldr r9, [r8], # [all...] |
H A D | ssi-fiq.S | 10 * r8 = bit 0-15: tx offset, bit 16-31: tx buffer size 59 and r10, r10, r8 /* r10: current buffer offset */ 76 lsr r11, r8, #16 /* r11: buffer size */ 78 lslgt r8, r11, #16 79 addle r8, #8
|
/linux-master/arch/powerpc/lib/ |
H A D | memcpy_64.S | 74 mr r8,r9 77 std r8,8(r3) 78 2: ldu r8,16(r4) 81 3: std r8,8(r3) 120 sld r8,r0,r10 124 # s1<< in r8, d0=(s0<<|s1>>) in r7, s3 in r0, s2 in r9, nix in r6 & r12 129 sld r8,r0,r10 136 or r12,r8,r12 138 sld r8,r0,r10 142 # d0=(s0<<|s1>>) in r12, s1<< in r6, s2>> in r7, s2<< in r8, s [all...] |
H A D | copy_32.S | 17 lwz r8,8(r4); \ 21 stw r8,8(r6); \ 29 lwz r8,8(r4); \ 37 stw r8,8(r6); \ 111 add r8,r7,r5 112 srwi r9,r8,LG_CACHELINE_BYTES 126 clrlwi r5,r8,32-LG_CACHELINE_BYTES 173 add r8,r4,r5 175 cmplw 1,r3,r8 187 andi. r8,r [all...] |
/linux-master/tools/testing/selftests/powerpc/copyloops/ |
H A D | memcpy_64.S | 74 mr r8,r9 77 std r8,8(r3) 78 2: ldu r8,16(r4) 81 3: std r8,8(r3) 120 sld r8,r0,r10 124 # s1<< in r8, d0=(s0<<|s1>>) in r7, s3 in r0, s2 in r9, nix in r6 & r12 129 sld r8,r0,r10 136 or r12,r8,r12 138 sld r8,r0,r10 142 # d0=(s0<<|s1>>) in r12, s1<< in r6, s2>> in r7, s2<< in r8, s [all...] |
/linux-master/arch/sh/lib/ |
H A D | copy_page.S | 22 * r8 --- from + PAGE_SIZE 28 mov.l r8,@-r15 33 mov r5,r8 37 add r0,r8 60 cmp/eq r11,r8 66 mov.l @r15+,r8 99 mov.l r8,@-r15 195 EX( mov.l @r5+,r8 ) 210 EX( mov.l r8,@(16,r4) ) 246 EX( mov.l @r5+,r8 ) [all...] |
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | hvCall.S | 41 std r8,STK_PARAM(R8)(r1); \ 53 ld r8,STACK_FRAME_MIN_SIZE+STK_PARAM(R8)(r1); \ 163 mr r7,r8 164 mr r8,r9 190 mr r7,r8 191 mr r8,r9 230 mr r7,r8 231 mr r8,r9 263 mr r7,r8 264 mr r8,r [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | processor.h | 54 unsigned long r7, r8, r9; \ 58 r8 = regs->ARM_r8; \ 65 regs->ARM_r8 = r8; \
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | ptrace.h | 59 unsigned long r8; member in struct:pt_regs
|
/linux-master/arch/arc/include/uapi/asm/ |
H A D | ptrace.h | 40 unsigned long r12, r11, r10, r9, r8, r7, r6, r5, r4, r3, r2, r1, r0; member in struct:user_regs_struct::__anon131
|
/linux-master/arch/arc/lib/ |
H A D | memcpy-archs-unaligned.S | 30 LOADX (r8, r1) 34 STOREX (r8, r3)
|
H A D | memcpy-archs.S | 66 LOADX (r8, r1) 70 STOREX (r8, r3) 109 ld.ab r8, [r1,4] 115 SHIFT_1 (r9, r8, 24) 117 SHIFT_2 (r5, r8, 8) 152 ld.ab r8, [r1,4] 158 SHIFT_1 (r9, r8, 16) 160 SHIFT_2 (r5, r8, 16) 192 ld.ab r8, [r1,4] 198 SHIFT_1 (r9, r8, [all...] |
/linux-master/arch/alpha/include/uapi/asm/ |
H A D | ptrace.h | 29 unsigned long r8; member in struct:pt_regs
|
/linux-master/arch/hexagon/include/uapi/asm/ |
H A D | user.h | 22 unsigned long r8; member in struct:user_regs_struct
|
/linux-master/arch/x86/lib/ |
H A D | memmove_64.S | 36 mov %rsi, %r8 37 add %rdx, %r8 38 cmp %rdi, %r8 70 movq 3*8(%rsi), %r8 76 movq %r8, 3*8(%rdi) 138 movq -4*8(%rsi), %r8 144 movq %r8, -4*8(%rdi) 162 movq -1*8(%rsi, %rdx), %r8 166 movq %r8, -1*8(%rdi, %rdx)
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_rmhandlers.S | 130 ld r8, 112+PPC_LR_STKOFF(r1) 142 mtsrr0 r8 163 ld r8,VCORE_LPCR(r5) 164 mtspr SPRN_LPCR,r8 548 addi r8, r5, VCORE_ENTRY_EXIT 549 21: lwarx r3, 0, r8 553 stwcx. r3, 0, r8 578 22: ld r8,VCORE_TB_OFFSET(r5) 579 cmpdi r8,0 581 std r8, VCORE_TB_OFFSET_APP [all...] |
/linux-master/arch/arm/kernel/ |
H A D | fiqasm.S | 30 ldmia r0!, {r8 - r12} 43 stmia r0!, {r8 - r12}
|
/linux-master/arch/powerpc/boot/ |
H A D | div64.S | 20 li r8,0 43 add r8,r8,r11 # and add the estimate to the accumulated 50 add r8,r8,r0 53 stw r8,4(r3) 77 rlwinm r8,r7,0,32 # t3 = (count < 32) ? 32 : 0 80 slw r7,r7,r8 # t2 = (count < 32) ? 0 : t2
|
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | subcore-asm.S | 45 mfspr r8, SPRN_PMCR 78 mtspr SPRN_PMCR, r8
|