/linux-master/arch/powerpc/lib/ |
H A D | xor_vmx_glue.c | 15 void xor_altivec_2(unsigned long bytes, unsigned long * __restrict p1, argument 20 __xor_altivec_2(bytes, p1, p2); 26 void xor_altivec_3(unsigned long bytes, unsigned long * __restrict p1, argument 32 __xor_altivec_3(bytes, p1, p2, p3); 38 void xor_altivec_4(unsigned long bytes, unsigned long * __restrict p1, argument 45 __xor_altivec_4(bytes, p1, p2, p3, p4); 51 void xor_altivec_5(unsigned long bytes, unsigned long * __restrict p1, argument 59 __xor_altivec_5(bytes, p1, p2, p3, p4, p5);
|
H A D | pmem.c | 16 unsigned long bytes = l1_dcache_bytes(); local 17 void *addr = (void *)(start & ~(bytes - 1)); 18 unsigned long size = stop - (unsigned long)addr + (bytes - 1); 21 for (i = 0; i < size >> shift; i++, addr += bytes) 28 unsigned long bytes = l1_dcache_bytes(); local 29 void *addr = (void *)(start & ~(bytes - 1)); 30 unsigned long size = stop - (unsigned long)addr + (bytes - 1); 33 for (i = 0; i < size >> shift; i++, addr += bytes)
|
/linux-master/arch/riscv/include/asm/ |
H A D | xor.h | 13 static void xor_vector_2(unsigned long bytes, unsigned long *__restrict p1, argument 17 xor_regs_2_(bytes, p1, p2); 21 static void xor_vector_3(unsigned long bytes, unsigned long *__restrict p1, argument 26 xor_regs_3_(bytes, p1, p2, p3); 30 static void xor_vector_4(unsigned long bytes, unsigned long *__restrict p1, argument 36 xor_regs_4_(bytes, p1, p2, p3, p4); 40 static void xor_vector_5(unsigned long bytes, unsigned long *__restrict p1, argument 47 xor_regs_5_(bytes, p1, p2, p3, p4, p5);
|
/linux-master/lib/crypto/ |
H A D | libchacha.c | 16 unsigned int bytes, int nrounds) 21 while (bytes >= CHACHA_BLOCK_SIZE) { 24 bytes -= CHACHA_BLOCK_SIZE; 28 if (bytes) { 30 crypto_xor_cpy(dst, src, stream, bytes); 15 chacha_crypt_generic(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, int nrounds) argument
|
/linux-master/arch/x86/crypto/ |
H A D | chacha_glue.c | 49 unsigned int bytes, int nrounds) 53 while (bytes >= CHACHA_BLOCK_SIZE * 8) { 54 chacha_8block_xor_avx512vl(state, dst, src, bytes, 56 bytes -= CHACHA_BLOCK_SIZE * 8; 61 if (bytes > CHACHA_BLOCK_SIZE * 4) { 62 chacha_8block_xor_avx512vl(state, dst, src, bytes, 64 state[12] += chacha_advance(bytes, 8); 67 if (bytes > CHACHA_BLOCK_SIZE * 2) { 68 chacha_4block_xor_avx512vl(state, dst, src, bytes, 70 state[12] += chacha_advance(bytes, 48 chacha_dosimd(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, int nrounds) argument 142 chacha_crypt_arch(u32 *state, u8 *dst, const u8 *src, unsigned int bytes, int nrounds) argument [all...] |
/linux-master/tools/include/linux/ |
H A D | overflow.h | 79 * Returns: number of bytes needed to represent the array or SIZE_MAX on 84 size_t bytes; local 86 if (check_mul_overflow(a, b, &bytes)) 89 return bytes; 101 * Returns: number of bytes needed to represent the array or SIZE_MAX on 106 size_t bytes; local 108 if (check_mul_overflow(a, b, &bytes)) 110 if (check_mul_overflow(bytes, c, &bytes)) 113 return bytes; 118 size_t bytes; local [all...] |
/linux-master/arch/hexagon/mm/ |
H A D | copy_user_template.S | 8 * W: width in bytes 19 p0 = cmp.gtu(bytes,#0) 25 p1 = cmp.gtu(bytes,#15) 32 loopcount = lsr(bytes,#3) 46 bytes -= asl(loopcount,#3) 58 p1 = cmp.gtu(bytes,#7) 63 loopcount = lsr(bytes,#2) 76 bytes -= asl(loopcount,#2) 84 p1 = cmp.gtu(bytes,#3) 89 loopcount = lsr(bytes,# [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | netcnt_prog.c | 37 percpu_cnt->bytes += skb->len; 44 __sync_fetch_and_add(&cnt->bytes, 45 percpu_cnt->bytes); 46 percpu_cnt->bytes = 0; 55 if (cnt->bytes + percpu_cnt->bytes - percpu_cnt->prev_bytes < dt) 63 percpu_cnt->prev_bytes = cnt->bytes;
|
H A D | test_sockmap_kern.h | 239 int *bytes, zero = 0, one = 1, two = 2, three = 3, four = 4, five = 5; local 242 bytes = bpf_map_lookup_elem(&sock_apply_bytes, &zero); 243 if (bytes) 244 bpf_msg_apply_bytes(msg, *bytes); 245 bytes = bpf_map_lookup_elem(&sock_cork_bytes, &zero); 246 if (bytes) 247 bpf_msg_cork_bytes(msg, *bytes); 270 int *bytes, *start, *end, *start_push, *end_push, *start_pop, *pop, *f; local 274 bytes = bpf_map_lookup_elem(&sock_apply_bytes, &zero); 275 if (bytes) 316 int ret = 0, *bytes, zero = 0; local 337 int ret = 0, *bytes, zero = 0; local 353 int *bytes, *start, *end, *start_push, *end_push, *start_pop, *pop; local [all...] |
/linux-master/fs/squashfs/ |
H A D | symlink.c | 43 int bytes, copied; local 51 * Skip index bytes into symlink metadata. 54 bytes = squashfs_read_metadata(sb, NULL, &block, &offset, 56 if (bytes < 0) { 65 * Read length bytes from symlink metadata. Squashfs_read_metadata 68 * squashfs_cache_get routine. As length bytes may overlap metadata 71 for (bytes = 0; bytes < length; offset = 0, bytes += copied) { 82 copied = squashfs_copy_data(pageaddr + bytes, entr [all...] |
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | mmio.c | 65 void *p_data, unsigned int bytes, bool read) 80 bytes); 83 bytes); 88 memcpy(p_data, pt, bytes); 90 memcpy(pt, p_data, bytes); 101 * @bytes: access data length 107 void *p_data, unsigned int bytes) 115 failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, true); 122 if (drm_WARN_ON(&i915->drm, bytes > 8)) 129 if (drm_WARN_ON(&i915->drm, bytes ! 64 failsafe_emulate_mmio_rw(struct intel_vgpu *vgpu, u64 pa, void *p_data, unsigned int bytes, bool read) argument 106 intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, u64 pa, void *p_data, unsigned int bytes) argument 181 intel_vgpu_emulate_mmio_write(struct intel_vgpu *vgpu, u64 pa, void *p_data, unsigned int bytes) argument [all...] |
/linux-master/include/linux/ |
H A D | uio.h | 145 * Total number of bytes covered by an iovec. 162 size_t bytes, struct iov_iter *i); 163 void iov_iter_advance(struct iov_iter *i, size_t bytes); 164 void iov_iter_revert(struct iov_iter *i, size_t bytes); 165 size_t fault_in_iov_iter_readable(const struct iov_iter *i, size_t bytes); 166 size_t fault_in_iov_iter_writeable(const struct iov_iter *i, size_t bytes); 168 size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, 170 size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, 173 size_t _copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i); 174 size_t _copy_from_iter(void *addr, size_t bytes, struc 177 copy_folio_to_iter(struct folio *folio, size_t offset, size_t bytes, struct iov_iter *i) argument 183 copy_folio_from_iter_atomic(struct folio *folio, size_t offset, size_t bytes, struct iov_iter *i) argument 193 copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i) argument 201 copy_from_iter(void *addr, size_t bytes, struct iov_iter *i) argument 209 copy_from_iter_full(void *addr, size_t bytes, struct iov_iter *i) argument 219 copy_from_iter_nocache(void *addr, size_t bytes, struct iov_iter *i) argument 227 copy_from_iter_full_nocache(void *addr, size_t bytes, struct iov_iter *i) argument [all...] |
/linux-master/lib/raid6/ |
H A D | neon.c | 31 size_t bytes, void **ptrs) \ 37 (unsigned long)bytes, ptrs); \ 42 size_t bytes, void **ptrs) \ 48 start, stop, (unsigned long)bytes, ptrs); \
|
H A D | recov_avx2.c | 16 static void raid6_2data_recov_avx2(int disks, size_t bytes, int faila, argument 37 raid6_call.gen_syndrome(disks, bytes, ptrs); 55 while (bytes) { 128 bytes -= 64; 175 bytes -= 32; 186 static void raid6_datap_recov_avx2(int disks, size_t bytes, int faila, argument 202 raid6_call.gen_syndrome(disks, bytes, ptrs); 215 while (bytes) { 263 bytes -= 64; 293 bytes [all...] |
H A D | recov.c | 19 static void raid6_2data_recov_intx1(int disks, size_t bytes, int faila, argument 40 raid6_call.gen_syndrome(disks, bytes, ptrs); 53 while ( bytes-- ) { 63 static void raid6_datap_recov_intx1(int disks, size_t bytes, int faila, argument 78 raid6_call.gen_syndrome(disks, bytes, ptrs); 88 while ( bytes-- ) { 107 void raid6_dual_recov(int disks, size_t bytes, int faila, int failb, void **ptrs) argument 118 raid6_call.gen_syndrome(disks, bytes, ptrs); 127 raid6_datap_recov(disks, bytes, faila, ptrs); 130 raid6_2data_recov(disks, bytes, fail [all...] |
/linux-master/include/crypto/ |
H A D | ghash.h | 21 u32 bytes; member in struct:ghash_desc_ctx
|
/linux-master/drivers/staging/media/atomisp/pci/ |
H A D | ia_css_firmware.h | 31 unsigned int bytes; /** length in bytes of firmware data */ member in struct:ia_css_fw
|
/linux-master/arch/s390/crypto/ |
H A D | ghash_s390.c | 27 u32 bytes; member in struct:ghash_desc_ctx 61 if (dctx->bytes) { 62 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); 64 n = min(srclen, dctx->bytes); 65 dctx->bytes -= n; 71 if (!dctx->bytes) { 85 dctx->bytes = GHASH_BLOCK_SIZE - srclen; 96 if (dctx->bytes) { 97 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); 99 memset(pos, 0, dctx->bytes); [all...] |
/linux-master/include/net/ |
H A D | netdev_queues.h | 9 u64 bytes; member in struct:netdev_queue_stats_rx 15 u64 bytes; member in struct:netdev_queue_stats_tx 142 * @bytes != 0, regardless of kernel config. 146 unsigned int pkts, unsigned int bytes) 149 netdev_tx_completed_queue(dev_queue, pkts, bytes); 150 else if (bytes) 158 * @bytes: number of bytes completed 167 * Reports completed pkts/bytes to BQL. 174 #define __netif_txq_completed_wake(txq, pkts, bytes, \ 145 netdev_txq_completed_mb(struct netdev_queue *dev_queue, unsigned int pkts, unsigned int bytes) argument [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | netcnt_common.h | 11 * It is about 128 bytes on x86_64 and 512 bytes on s390x, but allocate more to 26 __u64 bytes; member in struct:percpu_net_cnt::__anon5156 39 __u64 bytes; member in struct:net_cnt::__anon5157
|
/linux-master/arch/loongarch/lib/ |
H A D | xor_simd_glue.c | 15 void xor_##flavor##_2(unsigned long bytes, unsigned long * __restrict p1, \ 19 __xor_##flavor##_2(bytes, p1, p2); \ 25 void xor_##flavor##_3(unsigned long bytes, unsigned long * __restrict p1, \ 30 __xor_##flavor##_3(bytes, p1, p2, p3); \ 36 void xor_##flavor##_4(unsigned long bytes, unsigned long * __restrict p1, \ 42 __xor_##flavor##_4(bytes, p1, p2, p3, p4); \ 48 void xor_##flavor##_5(unsigned long bytes, unsigned long * __restrict p1, \ 55 __xor_##flavor##_5(bytes, p1, p2, p3, p4, p5); \
|
H A D | xor_template.c | 16 void XOR_FUNC_NAME(2)(unsigned long bytes, variable 20 unsigned long lines = bytes / LINE_WIDTH; 35 void XOR_FUNC_NAME(3)(unsigned long bytes, variable 40 unsigned long lines = bytes / LINE_WIDTH; 57 void XOR_FUNC_NAME(4)(unsigned long bytes, variable 63 unsigned long lines = bytes / LINE_WIDTH; 83 void XOR_FUNC_NAME(5)(unsigned long bytes, variable 90 unsigned long lines = bytes / LINE_WIDTH;
|
/linux-master/drivers/staging/media/atomisp/pci/hive_isp_css_common/host/ |
H A D | sp_local.h | 50 #define store_sp_var(var, data, bytes) \ 51 sp_dmem_store(SP0_ID, (unsigned int)sp_address_of(var), data, bytes) 65 #define store_sp_var_with_offset(var, offset, data, bytes) \ 67 offset, data, bytes) 69 #define load_sp_var(var, data, bytes) \ 70 sp_dmem_load(SP0_ID, (unsigned int)sp_address_of(var), data, bytes) 72 #define load_sp_var_with_offset(var, offset, data, bytes) \ 74 data, bytes)
|
/linux-master/drivers/staging/media/atomisp/pci/hmm/ |
H A D | hmm.c | 169 static ia_css_ptr __hmm_alloc(size_t bytes, enum hmm_bo_type type, argument 184 pgnr = size_to_pgnr_ceil(bytes); 207 dev_dbg(atomisp_dev, "pages: 0x%08x (%zu bytes), type: %d, vmalloc %p\n", 208 bo->start, bytes, type, vmalloc); 220 ia_css_ptr hmm_alloc(size_t bytes) argument 222 return __hmm_alloc(bytes, HMM_BO_PRIVATE, NULL); 225 ia_css_ptr hmm_create_from_vmalloc_buf(size_t bytes, void *vmalloc_addr) argument 227 return __hmm_alloc(bytes, HMM_BO_VMALLOC, vmalloc_addr); 279 unsigned int bytes) 292 while (bytes) { 278 load_and_flush_by_kmap(ia_css_ptr virt, void *data, unsigned int bytes) argument 322 load_and_flush(ia_css_ptr virt, void *data, unsigned int bytes) argument 357 hmm_load(ia_css_ptr virt, void *data, unsigned int bytes) argument 373 hmm_flush(ia_css_ptr virt, unsigned int bytes) argument 379 hmm_store(ia_css_ptr virt, const void *data, unsigned int bytes) argument 462 hmm_set(ia_css_ptr virt, int c, unsigned int bytes) argument [all...] |
/linux-master/arch/powerpc/crypto/ |
H A D | ghash.c | 43 int bytes; member in struct:p8_ghash_desc_ctx 50 dctx->bytes = 0; 124 if (dctx->bytes) { 125 if (dctx->bytes + srclen < GHASH_DIGEST_SIZE) { 126 memcpy(dctx->buffer + dctx->bytes, src, 128 dctx->bytes += srclen; 131 memcpy(dctx->buffer + dctx->bytes, src, 132 GHASH_DIGEST_SIZE - dctx->bytes); 136 src += GHASH_DIGEST_SIZE - dctx->bytes; 137 srclen -= GHASH_DIGEST_SIZE - dctx->bytes; [all...] |