/linux-master/arch/arm64/crypto/ |
H A D | aes-glue.c | 88 int rounds, int bytes, u8 const iv[]); 90 int rounds, int bytes, u8 const iv[]); 93 int rounds, int bytes, u8 ctr[]); 96 int rounds, int bytes, u8 ctr[], int byte_ctr); 99 int rounds, int bytes, u32 const rk2[], u8 iv[], 102 int rounds, int bytes, u32 const rk2[], u8 iv[], 468 * If given less than 16 bytes, we must copy the partial block 469 * into a temporary buffer of 16 bytes to avoid out of bounds 513 * If given less than 16 bytes, we must copy the partial block 514 * into a temporary buffer of 16 bytes t [all...] |
H A D | aes-neonbs-glue.c | 50 int rounds, int bytes, u8 ctr[]); 52 u32 const rk1[], int rounds, int bytes, 55 u32 const rk1[], int rounds, int bytes,
|
H A D | chacha-neon-glue.c | 37 int nrounds, int bytes); 43 int bytes, int nrounds) 45 while (bytes > 0) { 46 int l = min(bytes, CHACHA_BLOCK_SIZE * 5); 58 bytes -= l; 83 void chacha_crypt_arch(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, argument 86 if (!static_branch_likely(&have_neon) || bytes <= CHACHA_BLOCK_SIZE || 88 return chacha_crypt_generic(state, dst, src, bytes, nrounds); 91 unsigned int todo = min_t(unsigned int, bytes, SZ_4K); 97 bytes 42 chacha_doneon(u32 *state, u8 *dst, const u8 *src, int bytes, int nrounds) argument [all...] |
H A D | ghash-ce-glue.c | 58 asmlinkage void pmull_gcm_encrypt(int bytes, u8 dst[], const u8 src[], 61 asmlinkage int pmull_gcm_decrypt(int bytes, u8 dst[], const u8 src[],
|
H A D | poly1305-armv8.pl | 81 rev $r0,$r0 // flip bytes 254 rev $h0,$h0 // flip output bytes
|
H A D | poly1305-glue.c | 85 u32 bytes = min(len, POLY1305_BLOCK_SIZE - dctx->buflen); local 87 memcpy(dctx->buf + dctx->buflen, src, bytes); 88 src += bytes; 89 len -= bytes; 90 dctx->buflen += bytes; 129 u32 bytes = min(nbytes, POLY1305_BLOCK_SIZE - dctx->buflen); local 131 memcpy(dctx->buf + dctx->buflen, src, bytes); 132 src += bytes; 133 nbytes -= bytes; 134 dctx->buflen += bytes; [all...] |
H A D | polyval-ce-glue.c | 41 u32 bytes; member in struct:polyval_desc_ctx 110 if (dctx->bytes) { 111 n = min(srclen, dctx->bytes); 112 pos = dctx->buffer + POLYVAL_BLOCK_SIZE - dctx->bytes; 114 dctx->bytes -= n; 120 if (!dctx->bytes) 126 /* allow rescheduling every 4K bytes */ 134 dctx->bytes = POLYVAL_BLOCK_SIZE - srclen; 148 if (dctx->bytes) {
|
/linux-master/arch/arm64/include/asm/ |
H A D | xor.h | 19 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument 23 xor_block_inner_neon.do_2(bytes, p1, p2); 28 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, argument 33 xor_block_inner_neon.do_3(bytes, p1, p2, p3); 38 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, argument 44 xor_block_inner_neon.do_4(bytes, p1, p2, p3, p4); 49 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, argument 56 xor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5);
|
/linux-master/arch/arm64/lib/ |
H A D | xor-neon.c | 13 static void xor_arm64_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument 20 long lines = bytes / (sizeof(uint64x2_t) * 4); 40 static void xor_arm64_neon_3(unsigned long bytes, unsigned long * __restrict p1, argument 49 long lines = bytes / (sizeof(uint64x2_t) * 4); 76 static void xor_arm64_neon_4(unsigned long bytes, unsigned long * __restrict p1, argument 87 long lines = bytes / (sizeof(uint64x2_t) * 4); 121 static void xor_arm64_neon_5(unsigned long bytes, unsigned long * __restrict p1, argument 134 long lines = bytes / (sizeof(uint64x2_t) * 4); 194 static void xor_arm64_eor3_3(unsigned long bytes, argument 204 long lines = bytes / (sizeo 229 xor_arm64_eor3_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) argument 273 xor_arm64_eor3_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) argument [all...] |
/linux-master/arch/csky/mm/ |
H A D | syscache.c | 11 unsigned long, bytes, 18 (unsigned long)addr + bytes); 25 (unsigned long)addr + bytes);
|
/linux-master/arch/loongarch/include/asm/ |
H A D | xor_simd.h | 9 void xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1, 11 void xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1, 13 void xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1, 16 void xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1, 22 void xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1, 24 void xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1, 26 void xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1, 29 void xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
|
/linux-master/arch/loongarch/lib/ |
H A D | xor_simd.h | 13 void __xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1, 15 void __xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1, 17 void __xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1, 20 void __xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1, 26 void __xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1, 28 void __xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1, 30 void __xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1, 33 void __xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
|
H A D | xor_simd_glue.c | 15 void xor_##flavor##_2(unsigned long bytes, unsigned long * __restrict p1, \ 19 __xor_##flavor##_2(bytes, p1, p2); \ 25 void xor_##flavor##_3(unsigned long bytes, unsigned long * __restrict p1, \ 30 __xor_##flavor##_3(bytes, p1, p2, p3); \ 36 void xor_##flavor##_4(unsigned long bytes, unsigned long * __restrict p1, \ 42 __xor_##flavor##_4(bytes, p1, p2, p3, p4); \ 48 void xor_##flavor##_5(unsigned long bytes, unsigned long * __restrict p1, \ 55 __xor_##flavor##_5(bytes, p1, p2, p3, p4, p5); \
|
H A D | xor_template.c | 16 void XOR_FUNC_NAME(2)(unsigned long bytes, variable 20 unsigned long lines = bytes / LINE_WIDTH; 35 void XOR_FUNC_NAME(3)(unsigned long bytes, variable 40 unsigned long lines = bytes / LINE_WIDTH; 57 void XOR_FUNC_NAME(4)(unsigned long bytes, variable 63 unsigned long lines = bytes / LINE_WIDTH; 83 void XOR_FUNC_NAME(5)(unsigned long bytes, variable 90 unsigned long lines = bytes / LINE_WIDTH;
|
/linux-master/arch/m68k/fpsp040/ |
H A D | skeleton.S | 395 | d0 - number of bytes to write (maximum count is 12) 440 | d0 - number of bytes to read (maximum count is 12)
|
/linux-master/arch/m68k/ifpsp060/ |
H A D | os.S | 81 | d0 - number of bytes to write 112 | d0 - number of bytes to read
|
/linux-master/arch/m68k/include/asm/ |
H A D | floppy.h | 192 static int bytes=0; local 202 bytes = virtual_dma_count; 238 pr_info("count=%x, residue=%x calls=%d bytes=%d dma_wait=%d\n", 239 virtual_dma_count, virtual_dma_residue, calls, bytes,
|
H A D | sun3xflop.h | 123 static int bytes=0; local 135 bytes = virtual_dma_count; 184 pr_info("count=%x, residue=%x calls=%d bytes=%x dma_wait=%d\n", 185 virtual_dma_count, virtual_dma_residue, calls, bytes,
|
/linux-master/arch/mips/alchemy/devboards/ |
H A D | pm.c | 151 size_t bytes) 187 bytes = -ENOENT; 189 return bytes; 148 db1x_pmattr_store(struct kobject *kobj, struct kobj_attribute *attr, const char *instr, size_t bytes) argument
|
/linux-master/arch/mips/cavium-octeon/ |
H A D | octeon-memcpy.S | 28 * memcpy copies len bytes from src to dst and sets v0 to dst. 35 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to 36 * the number of uncopied bytes due to an exception caused by a read or write. 64 * (2) is met by incrementing src by the number of bytes copied 145 * the number of uncopied bytes. 182 # This is where we loop if there is more than 128 bytes left 387 * Copy bytes from src until faulting load address (or until a 407 SUB len, AT, t0 # len number of uncopied bytes
|
/linux-master/arch/mips/crypto/ |
H A D | chacha-glue.c | 17 unsigned int bytes, int nrounds);
|
H A D | poly1305-glue.c | 76 u32 bytes = min(len, POLY1305_BLOCK_SIZE - dctx->buflen); local 78 memcpy(dctx->buf + dctx->buflen, src, bytes); 79 src += bytes; 80 len -= bytes; 81 dctx->buflen += bytes; 106 u32 bytes = min(nbytes, POLY1305_BLOCK_SIZE - dctx->buflen); local 108 memcpy(dctx->buf + dctx->buflen, src, bytes); 109 src += bytes; 110 nbytes -= bytes; 111 dctx->buflen += bytes; [all...] |
/linux-master/arch/mips/lib/ |
H A D | csum_partial.S | 112 bnez t8, .Lsmall_csumcpy /* < 8 bytes to copy */ 235 /* unknown src alignment and < 8 bytes to go */ 505 * len == the number of bytes left to copy < 8*NBYTES 552 * src and dst are aligned, need to copy rem bytes (rem < NBYTES) 583 * Copy enough bytes to align dst 590 SUB t2, t2, t1 # t2 = number of bytes copied
|
H A D | memcpy.S | 47 * memcpy copies len bytes from src to dst and sets v0 to dst. 54 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to 55 * the number of uncopied bytes due to an exception caused by a read or write. 83 * (2) is met by incrementing src by the number of bytes copied 350 * len == rem == the number of bytes left to copy < 8*NBYTES 394 * src and dst are aligned, need to copy rem bytes (rem < NBYTES) 422 * Copy enough bytes to align dst 429 SUB t2, t2, t1 # t2 = number of bytes copied 546 * Copy bytes from src until faulting load address (or until a 571 SUB len, AT, t0 # len number of uncopied bytes [all...] |
/linux-master/arch/mips/mm/ |
H A D | cache.c | 89 SYSCALL_DEFINE3(cacheflush, unsigned long, addr, unsigned long, bytes, 92 if (bytes == 0) 94 if (!access_ok((void __user *) addr, bytes)) 97 __flush_icache_user_range(addr, addr + bytes);
|