Searched refs:bytes (Results 126 - 150 of 1671) sorted by path

1234567891011>>

/linux-master/arch/arm64/crypto/
H A Daes-glue.c88 int rounds, int bytes, u8 const iv[]);
90 int rounds, int bytes, u8 const iv[]);
93 int rounds, int bytes, u8 ctr[]);
96 int rounds, int bytes, u8 ctr[], int byte_ctr);
99 int rounds, int bytes, u32 const rk2[], u8 iv[],
102 int rounds, int bytes, u32 const rk2[], u8 iv[],
468 * If given less than 16 bytes, we must copy the partial block
469 * into a temporary buffer of 16 bytes to avoid out of bounds
513 * If given less than 16 bytes, we must copy the partial block
514 * into a temporary buffer of 16 bytes t
[all...]
H A Daes-neonbs-glue.c50 int rounds, int bytes, u8 ctr[]);
52 u32 const rk1[], int rounds, int bytes,
55 u32 const rk1[], int rounds, int bytes,
H A Dchacha-neon-glue.c37 int nrounds, int bytes);
43 int bytes, int nrounds)
45 while (bytes > 0) {
46 int l = min(bytes, CHACHA_BLOCK_SIZE * 5);
58 bytes -= l;
83 void chacha_crypt_arch(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, argument
86 if (!static_branch_likely(&have_neon) || bytes <= CHACHA_BLOCK_SIZE ||
88 return chacha_crypt_generic(state, dst, src, bytes, nrounds);
91 unsigned int todo = min_t(unsigned int, bytes, SZ_4K);
97 bytes
42 chacha_doneon(u32 *state, u8 *dst, const u8 *src, int bytes, int nrounds) argument
[all...]
H A Dghash-ce-glue.c58 asmlinkage void pmull_gcm_encrypt(int bytes, u8 dst[], const u8 src[],
61 asmlinkage int pmull_gcm_decrypt(int bytes, u8 dst[], const u8 src[],
H A Dpoly1305-armv8.pl81 rev $r0,$r0 // flip bytes
254 rev $h0,$h0 // flip output bytes
H A Dpoly1305-glue.c85 u32 bytes = min(len, POLY1305_BLOCK_SIZE - dctx->buflen); local
87 memcpy(dctx->buf + dctx->buflen, src, bytes);
88 src += bytes;
89 len -= bytes;
90 dctx->buflen += bytes;
129 u32 bytes = min(nbytes, POLY1305_BLOCK_SIZE - dctx->buflen); local
131 memcpy(dctx->buf + dctx->buflen, src, bytes);
132 src += bytes;
133 nbytes -= bytes;
134 dctx->buflen += bytes;
[all...]
H A Dpolyval-ce-glue.c41 u32 bytes; member in struct:polyval_desc_ctx
110 if (dctx->bytes) {
111 n = min(srclen, dctx->bytes);
112 pos = dctx->buffer + POLYVAL_BLOCK_SIZE - dctx->bytes;
114 dctx->bytes -= n;
120 if (!dctx->bytes)
126 /* allow rescheduling every 4K bytes */
134 dctx->bytes = POLYVAL_BLOCK_SIZE - srclen;
148 if (dctx->bytes) {
/linux-master/arch/arm64/include/asm/
H A Dxor.h19 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument
23 xor_block_inner_neon.do_2(bytes, p1, p2);
28 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, argument
33 xor_block_inner_neon.do_3(bytes, p1, p2, p3);
38 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, argument
44 xor_block_inner_neon.do_4(bytes, p1, p2, p3, p4);
49 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, argument
56 xor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5);
/linux-master/arch/arm64/lib/
H A Dxor-neon.c13 static void xor_arm64_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument
20 long lines = bytes / (sizeof(uint64x2_t) * 4);
40 static void xor_arm64_neon_3(unsigned long bytes, unsigned long * __restrict p1, argument
49 long lines = bytes / (sizeof(uint64x2_t) * 4);
76 static void xor_arm64_neon_4(unsigned long bytes, unsigned long * __restrict p1, argument
87 long lines = bytes / (sizeof(uint64x2_t) * 4);
121 static void xor_arm64_neon_5(unsigned long bytes, unsigned long * __restrict p1, argument
134 long lines = bytes / (sizeof(uint64x2_t) * 4);
194 static void xor_arm64_eor3_3(unsigned long bytes, argument
204 long lines = bytes / (sizeo
229 xor_arm64_eor3_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) argument
273 xor_arm64_eor3_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) argument
[all...]
/linux-master/arch/csky/mm/
H A Dsyscache.c11 unsigned long, bytes,
18 (unsigned long)addr + bytes);
25 (unsigned long)addr + bytes);
/linux-master/arch/loongarch/include/asm/
H A Dxor_simd.h9 void xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1,
11 void xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1,
13 void xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1,
16 void xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1,
22 void xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1,
24 void xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1,
26 void xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1,
29 void xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
/linux-master/arch/loongarch/lib/
H A Dxor_simd.h13 void __xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1,
15 void __xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1,
17 void __xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1,
20 void __xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1,
26 void __xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1,
28 void __xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1,
30 void __xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1,
33 void __xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
H A Dxor_simd_glue.c15 void xor_##flavor##_2(unsigned long bytes, unsigned long * __restrict p1, \
19 __xor_##flavor##_2(bytes, p1, p2); \
25 void xor_##flavor##_3(unsigned long bytes, unsigned long * __restrict p1, \
30 __xor_##flavor##_3(bytes, p1, p2, p3); \
36 void xor_##flavor##_4(unsigned long bytes, unsigned long * __restrict p1, \
42 __xor_##flavor##_4(bytes, p1, p2, p3, p4); \
48 void xor_##flavor##_5(unsigned long bytes, unsigned long * __restrict p1, \
55 __xor_##flavor##_5(bytes, p1, p2, p3, p4, p5); \
H A Dxor_template.c16 void XOR_FUNC_NAME(2)(unsigned long bytes, variable
20 unsigned long lines = bytes / LINE_WIDTH;
35 void XOR_FUNC_NAME(3)(unsigned long bytes, variable
40 unsigned long lines = bytes / LINE_WIDTH;
57 void XOR_FUNC_NAME(4)(unsigned long bytes, variable
63 unsigned long lines = bytes / LINE_WIDTH;
83 void XOR_FUNC_NAME(5)(unsigned long bytes, variable
90 unsigned long lines = bytes / LINE_WIDTH;
/linux-master/arch/m68k/fpsp040/
H A Dskeleton.S395 | d0 - number of bytes to write (maximum count is 12)
440 | d0 - number of bytes to read (maximum count is 12)
/linux-master/arch/m68k/ifpsp060/
H A Dos.S81 | d0 - number of bytes to write
112 | d0 - number of bytes to read
/linux-master/arch/m68k/include/asm/
H A Dfloppy.h192 static int bytes=0; local
202 bytes = virtual_dma_count;
238 pr_info("count=%x, residue=%x calls=%d bytes=%d dma_wait=%d\n",
239 virtual_dma_count, virtual_dma_residue, calls, bytes,
H A Dsun3xflop.h123 static int bytes=0; local
135 bytes = virtual_dma_count;
184 pr_info("count=%x, residue=%x calls=%d bytes=%x dma_wait=%d\n",
185 virtual_dma_count, virtual_dma_residue, calls, bytes,
/linux-master/arch/mips/alchemy/devboards/
H A Dpm.c151 size_t bytes)
187 bytes = -ENOENT;
189 return bytes;
148 db1x_pmattr_store(struct kobject *kobj, struct kobj_attribute *attr, const char *instr, size_t bytes) argument
/linux-master/arch/mips/cavium-octeon/
H A Docteon-memcpy.S28 * memcpy copies len bytes from src to dst and sets v0 to dst.
35 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to
36 * the number of uncopied bytes due to an exception caused by a read or write.
64 * (2) is met by incrementing src by the number of bytes copied
145 * the number of uncopied bytes.
182 # This is where we loop if there is more than 128 bytes left
387 * Copy bytes from src until faulting load address (or until a
407 SUB len, AT, t0 # len number of uncopied bytes
/linux-master/arch/mips/crypto/
H A Dchacha-glue.c17 unsigned int bytes, int nrounds);
H A Dpoly1305-glue.c76 u32 bytes = min(len, POLY1305_BLOCK_SIZE - dctx->buflen); local
78 memcpy(dctx->buf + dctx->buflen, src, bytes);
79 src += bytes;
80 len -= bytes;
81 dctx->buflen += bytes;
106 u32 bytes = min(nbytes, POLY1305_BLOCK_SIZE - dctx->buflen); local
108 memcpy(dctx->buf + dctx->buflen, src, bytes);
109 src += bytes;
110 nbytes -= bytes;
111 dctx->buflen += bytes;
[all...]
/linux-master/arch/mips/lib/
H A Dcsum_partial.S112 bnez t8, .Lsmall_csumcpy /* < 8 bytes to copy */
235 /* unknown src alignment and < 8 bytes to go */
505 * len == the number of bytes left to copy < 8*NBYTES
552 * src and dst are aligned, need to copy rem bytes (rem < NBYTES)
583 * Copy enough bytes to align dst
590 SUB t2, t2, t1 # t2 = number of bytes copied
H A Dmemcpy.S47 * memcpy copies len bytes from src to dst and sets v0 to dst.
54 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to
55 * the number of uncopied bytes due to an exception caused by a read or write.
83 * (2) is met by incrementing src by the number of bytes copied
350 * len == rem == the number of bytes left to copy < 8*NBYTES
394 * src and dst are aligned, need to copy rem bytes (rem < NBYTES)
422 * Copy enough bytes to align dst
429 SUB t2, t2, t1 # t2 = number of bytes copied
546 * Copy bytes from src until faulting load address (or until a
571 SUB len, AT, t0 # len number of uncopied bytes
[all...]
/linux-master/arch/mips/mm/
H A Dcache.c89 SYSCALL_DEFINE3(cacheflush, unsigned long, addr, unsigned long, bytes,
92 if (bytes == 0)
94 if (!access_ok((void __user *) addr, bytes))
97 __flush_icache_user_range(addr, addr + bytes);

Completed in 358 milliseconds

1234567891011>>