/linux-master/arch/alpha/lib/ |
H A D | ev67-strchr.S | 54 xor t0, a1, t1 # E : make bytes == c zero
|
H A D | ev67-strrchr.S | 10 * - uses bcmpge to compare 8 bytes in parallel 57 xor t0, a1, t2 # E : make bytes == c zero 86 subq t4, 1, t5 # E : build a mask of the bytes up to...
|
H A D | memchr.S | 23 - uses cmpbge to compare 8 bytes in parallel 67 ldq_u $6, -1($5) #-e1 : eight or less bytes to search 72 # Deal with the case where at most 8 bytes remain to be searched 103 # Deal with the case where $18 > 8 bytes remain to be 157 $final: subq $5, $0, $18 # .. e1 : $18 <- number of bytes left to do
|
H A D | strchr.S | 35 xor t0, a1, t1 # .. e1 : make bytes == c zero
|
H A D | strrchr.S | 36 xor t0, a1, t2 # e0 : make bytes == c zero 58 subq t4, 1, t5 # e0 : build a mask of the bytes up to...
|
/linux-master/arch/arc/include/asm/ |
H A D | arcregs.h | 152 #define TO_KB(bytes) ((bytes) >> 10) 153 #define TO_MB(bytes) (TO_KB(bytes) >> 10)
|
/linux-master/arch/arm/ |
H A D | Makefile | 144 KBUILD_CFLAGS +=$(CFLAGS_ABI) $(CFLAGS_ISA) $(arch-y) $(tune-y) $(call cc-option,-mshort-load-bytes,$(call cc-option,-malignment-traps,)) -msoft-float -Uarm
|
/linux-master/arch/arm/boot/bootp/ |
H A D | init.S | 63 move: ldmia r4!, {r7 - r10} @ move 32-bytes at a time
|
/linux-master/arch/arm/boot/compressed/ |
H A D | efi-header.S | 19 @ PE/COFF signature "MZ" in the first two bytes, so the kernel 41 @ PE/COFF offset, and the "MZ" bytes at offset 0x0.
|
H A D | head.S | 195 * jump 32 bytes into the image to skip the a.out header. 196 * with these 8 nops filling exactly 32 bytes, things still 507 bic r9, r9, #31 @ ... of 32 bytes 700 mov \reg, #4 @ bytes per word 1313 mov r11, r11, lsl r3 @ cache line size in bytes
|
/linux-master/arch/arm/crypto/ |
H A D | aes-ce-core.S | 291 * int rounds, int bytes, u8 const iv[]) 293 * int rounds, int bytes, u8 const iv[]) 470 * int bytes, u8 iv[], u32 const rk2[], int first) 472 * int bytes, u8 iv[], u32 const rk2[], int first) 569 add r4, r4, #16 @ # bytes in final block 597 /* subtract 16 bytes if we are doing CTS */ 655 add r4, r4, #16 @ # bytes in final block
|
H A D | aes-ce-glue.c | 39 int rounds, int bytes, u8 const iv[]); 41 int rounds, int bytes, u8 const iv[]); 47 int rounds, int bytes, u8 iv[], 50 int rounds, int bytes, u8 iv[],
|
H A D | aes-neonbs-glue.c | 248 int bytes = walk.nbytes; local 250 if (unlikely(bytes < AES_BLOCK_SIZE)) 251 src = dst = memcpy(buf + sizeof(buf) - bytes, 252 src, bytes); 254 bytes &= ~(8 * AES_BLOCK_SIZE - 1); 257 aesbs_ctr_encrypt(dst, src, ctx->rk, ctx->rounds, bytes, walk.iv); 260 if (unlikely(bytes < AES_BLOCK_SIZE)) 262 buf + sizeof(buf) - bytes, bytes); 264 err = skcipher_walk_done(&walk, walk.nbytes - bytes); [all...] |
H A D | chacha-glue.c | 30 asmlinkage void chacha_doarm(u8 *dst, const u8 *src, unsigned int bytes, 41 unsigned int bytes, int nrounds) 45 while (bytes > CHACHA_BLOCK_SIZE) { 46 unsigned int l = min(bytes, CHACHA_BLOCK_SIZE * 4U); 49 bytes -= l; 54 if (bytes) { 58 if (bytes != CHACHA_BLOCK_SIZE) 59 s = d = memcpy(buf, src, bytes); 62 memcpy(dst, buf, bytes); 85 void chacha_crypt_arch(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, argument 40 chacha_doneon(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, int nrounds) argument [all...] |
H A D | ghash-ce-glue.c | 373 void pmull_gcm_decrypt(int bytes, u64 dg[], const char *src, 377 int pmull_gcm_dec_final(int bytes, u64 dg[], char *tag,
|
H A D | poly1305-glue.c | 89 u32 bytes = min(len, POLY1305_BLOCK_SIZE - dctx->buflen); local 91 memcpy(dctx->buf + dctx->buflen, src, bytes); 92 src += bytes; 93 len -= bytes; 94 dctx->buflen += bytes; 146 u32 bytes = min(nbytes, POLY1305_BLOCK_SIZE - dctx->buflen); local 148 memcpy(dctx->buf + dctx->buflen, src, bytes); 149 src += bytes; 150 nbytes -= bytes; 151 dctx->buflen += bytes; [all...] |
H A D | sha1-armv4-large.S | 25 @ impl size in bytes comp cycles[*] measured performance 390 bcs .L_done @ [+((12+3)*5+2)*4], spare 300 bytes 486 b .L_20_39_or_60_79 @ [+4], spare 300 bytes
|
/linux-master/arch/arm/include/asm/ |
H A D | xor.h | 47 xor_arm4regs_2(unsigned long bytes, unsigned long * __restrict p1, argument 50 unsigned int lines = bytes / sizeof(unsigned long) / 4; 68 xor_arm4regs_3(unsigned long bytes, unsigned long * __restrict p1, argument 72 unsigned int lines = bytes / sizeof(unsigned long) / 4; 91 xor_arm4regs_4(unsigned long bytes, unsigned long * __restrict p1, argument 96 unsigned int lines = bytes / sizeof(unsigned long) / 2; 112 xor_arm4regs_5(unsigned long bytes, unsigned long * __restrict p1, argument 118 unsigned int lines = bytes / sizeof(unsigned long) / 2; 156 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument 160 xor_arm4regs_2(bytes, p 169 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3) argument 183 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) argument 198 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) argument [all...] |
/linux-master/arch/arm/kernel/ |
H A D | entry-header.S | 23 @ the addition of 8 bytes for storing syscall args 5 and 6.
|
H A D | unwind.c | 313 unsigned long bytes = 0; local 321 * Note: This decodes a maximum of 4 bytes to output 28 bits data where 327 result |= (insn & 0x7f) << (bytes * 7); 328 bytes++; 329 } while (!!(insn & 0x80) && (bytes != sizeof(result)));
|
/linux-master/arch/arm/lib/ |
H A D | csumpartialcopygeneric.S | 28 * we have >= 8 bytes here, so we don't need to check 52 * Handle 0 to 7 bytes, with any alignment of source and 92 blo .Lless8 @ 8 bytes to copy. 100 * that we must have more than 4 bytes to copy. Note
|
H A D | memset.S | 46 stmiage ip!, {r1, r3, r8, lr} @ 64 bytes at a time. 51 ldmfdeq sp!, {r8, pc} @ Now <64 bytes to go. 115 * When we get here, we've got less than 4 bytes to set. We 126 blt 5b @ 1 bytes to align with?
|
/linux-master/arch/arm/mach-imx/ |
H A D | mmdc.c | 90 PMU_EVENT_ATTR_STRING(read-bytes, mmdc_pmu_read_bytes, "event=0x04") 91 PMU_EVENT_ATTR_STRING(read-bytes.unit, mmdc_pmu_read_bytes_unit, "MB"); 92 PMU_EVENT_ATTR_STRING(read-bytes.scale, mmdc_pmu_read_bytes_scale, "0.000001"); 93 PMU_EVENT_ATTR_STRING(write-bytes, mmdc_pmu_write_bytes, "event=0x05") 94 PMU_EVENT_ATTR_STRING(write-bytes.unit, mmdc_pmu_write_bytes_unit, "MB"); 95 PMU_EVENT_ATTR_STRING(write-bytes.scale, mmdc_pmu_write_bytes_scale, "0.000001");
|
/linux-master/arch/arm/mm/ |
H A D | cache-v4wb.S | 89 sub r1, r2, #512 @ only 512 bytes
|
H A D | proc-macros.S | 84 mov \reg, #4 @ bytes per word 101 mov \reg, #4 @ bytes per word
|