/linux-master/include/uapi/linux/netfilter/ |
H A D | xt_sctp.h | 39 #define bytes(type) (sizeof(type) * 8) macro 43 (chunkmap)[type / bytes(__u32)] |= \ 44 1u << (type % bytes(__u32)); \ 49 (chunkmap)[type / bytes(__u32)] &= \ 50 ~(1u << (type % bytes(__u32))); \ 55 ((chunkmap)[type / bytes (__u32)] & \ 56 (1u << (type % bytes (__u32)))) ? 1: 0; \
|
/linux-master/drivers/char/hw_random/ |
H A D | arm_smccc_trng.c | 37 size_t bytes) 41 if (bytes == 0) 44 chunk = min(bytes, sizeof(long)); 47 if (copied >= bytes) 50 chunk = min((bytes - copied), sizeof(long)); 53 if (copied >= bytes) 56 chunk = min((bytes - copied), sizeof(long)); 36 copy_from_registers(char *buf, struct arm_smccc_res *res, size_t bytes) argument
|
/linux-master/net/netfilter/ |
H A D | xt_connbytes.c | 16 MODULE_DESCRIPTION("Xtables: Number of packets/bytes per connection matching"); 27 u_int64_t bytes = 0; local 59 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 62 what = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); 65 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 66 what += atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); 73 bytes = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 77 bytes = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); [all...] |
/linux-master/fs/btrfs/ |
H A D | space-info.h | 99 u64 total_bytes; /* total bytes in the space, 101 u64 bytes_used; /* total bytes used, 103 u64 bytes_pinned; /* total bytes pinned, will be freed when the 105 u64 bytes_reserved; /* total bytes the allocator has reserved for 107 u64 bytes_may_use; /* number of bytes that may be used for 109 u64 bytes_readonly; /* total bytes that are read only */ 110 u64 bytes_zone_unusable; /* total bytes that are unusable until 116 /* Chunk size in bytes */ 138 u64 disk_used; /* total bytes used on disk */ 139 u64 disk_total; /* total bytes o 171 u64 bytes; member in struct:reserve_ticket [all...] |
/linux-master/arch/riscv/crypto/ |
H A D | ghash-riscv64-glue.c | 30 u32 bytes; member in struct:riscv64_ghash_desc_ctx 83 if (dctx->bytes) { 84 if (dctx->bytes + srclen < GHASH_BLOCK_SIZE) { 85 memcpy(dctx->buffer + dctx->bytes, src, srclen); 86 dctx->bytes += srclen; 89 memcpy(dctx->buffer + dctx->bytes, src, 90 GHASH_BLOCK_SIZE - dctx->bytes); 93 src += GHASH_BLOCK_SIZE - dctx->bytes; 94 srclen -= GHASH_BLOCK_SIZE - dctx->bytes; 95 dctx->bytes [all...] |
/linux-master/tools/lib/ |
H A D | string.c | 172 static void *check_bytes8(const u8 *start, u8 value, unsigned int bytes) argument 174 while (bytes) { 178 bytes--; 187 * @bytes: The size of the area. 192 void *memchr_inv(const void *start, int c, size_t bytes) argument 198 if (bytes <= 16) 199 return check_bytes8(start, value, bytes); 215 bytes -= prefix; 218 words = bytes / 8; 227 return check_bytes8(start, value, bytes [all...] |
/linux-master/include/linux/ |
H A D | uio.h | 145 * Total number of bytes covered by an iovec. 162 size_t bytes, struct iov_iter *i); 163 void iov_iter_advance(struct iov_iter *i, size_t bytes); 164 void iov_iter_revert(struct iov_iter *i, size_t bytes); 165 size_t fault_in_iov_iter_readable(const struct iov_iter *i, size_t bytes); 166 size_t fault_in_iov_iter_writeable(const struct iov_iter *i, size_t bytes); 168 size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, 170 size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, 173 size_t _copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i); 174 size_t _copy_from_iter(void *addr, size_t bytes, struc 177 copy_folio_to_iter(struct folio *folio, size_t offset, size_t bytes, struct iov_iter *i) argument 183 copy_folio_from_iter_atomic(struct folio *folio, size_t offset, size_t bytes, struct iov_iter *i) argument 193 copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i) argument 201 copy_from_iter(void *addr, size_t bytes, struct iov_iter *i) argument 209 copy_to_iter_full(const void *addr, size_t bytes, struct iov_iter *i) argument 219 copy_from_iter_full(void *addr, size_t bytes, struct iov_iter *i) argument 229 copy_from_iter_nocache(void *addr, size_t bytes, struct iov_iter *i) argument 237 copy_from_iter_full_nocache(void *addr, size_t bytes, struct iov_iter *i) argument [all...] |
H A D | alloc_tag.h | 19 u64 bytes; member in struct:alloc_tag_counters 59 s64 bytes; member in struct:codetag_bytes 111 v.bytes += counter->bytes; 159 static inline void alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag, size_t bytes) argument 166 this_cpu_add(tag->counters->bytes, bytes); 169 static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) argument 184 this_cpu_sub(tag->counters->bytes, bytes); 196 alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag, size_t bytes) argument 198 alloc_tag_sub(union codetag_ref *ref, size_t bytes) argument [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | xor.h | 47 xor_arm4regs_2(unsigned long bytes, unsigned long * __restrict p1, argument 50 unsigned int lines = bytes / sizeof(unsigned long) / 4; 68 xor_arm4regs_3(unsigned long bytes, unsigned long * __restrict p1, argument 72 unsigned int lines = bytes / sizeof(unsigned long) / 4; 91 xor_arm4regs_4(unsigned long bytes, unsigned long * __restrict p1, argument 96 unsigned int lines = bytes / sizeof(unsigned long) / 2; 112 xor_arm4regs_5(unsigned long bytes, unsigned long * __restrict p1, argument 118 unsigned int lines = bytes / sizeof(unsigned long) / 2; 156 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument 160 xor_arm4regs_2(bytes, p 169 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3) argument 183 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) argument 198 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) argument [all...] |
/linux-master/arch/x86/kvm/mmu/ |
H A D | page_track.h | 30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes); 42 const u8 *new, int bytes) { } 51 const u8 *new, int bytes) 53 __kvm_page_track_write(vcpu->kvm, gpa, new, bytes); 55 kvm_mmu_track_write(vcpu, gpa, new, bytes); 41 __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes) argument 50 kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, int bytes) argument
|
/linux-master/drivers/soc/tegra/fuse/ |
H A D | fuse-tegra30.c | 140 .bytes = 4, 146 .bytes = 4, 152 .bytes = 4, 158 .bytes = 4, 164 .bytes = 4, 170 .bytes = 4, 176 .bytes = 4, 182 .bytes = 4, 188 .bytes = 4, 194 .bytes [all...] |
/linux-master/crypto/ |
H A D | nhpoly1305.c | 88 * 1024 bytes, except possibly the final one which is taken over a multiple of 89 * 16 bytes up to 1024. Also, in the case where data is passed in misaligned 97 unsigned int bytes; local 101 bytes = min_t(unsigned int, srclen, NH_MESSAGE_BYTES); 102 nh_fn(key->nh_key, src, bytes, state->nh_hash); 103 state->nh_remaining = NH_MESSAGE_BYTES - bytes; 111 bytes = min(srclen, state->nh_remaining); 112 nh_fn(&key->nh_key[pos / 4], src, bytes, tmp_hash); 116 state->nh_remaining -= bytes; 120 src += bytes; 161 unsigned int bytes; local [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grukservices.h | 45 * Message queue size is the total number of bytes allocated 52 * bytes size of message queue in bytes 62 void *p, unsigned int bytes, int nasid, int vector, int apicid); 74 * bytes size of message in bytes 82 void *mesg, unsigned int bytes); 141 * bytes number of bytes to copy 148 unsigned int bytes); [all...] |
/linux-master/include/rdma/ |
H A D | iba.h | 46 * aligned, everything larger has a max alignment of 4 bytes. 71 /* FIXME: A set should always set the entire field, meaning we should zero the trailing bytes */ 73 bytes) \ 76 WARN_ON(bytes * 8 > num_bits); \ 77 if (in && bytes) \ 80 _in_ptr, bytes); \ 82 #define IBA_SET_MEM(field, ptr, in, bytes) _IBA_SET_MEM(field, ptr, in, bytes) 94 bytes) \ 97 WARN_ON(bytes * [all...] |
/linux-master/drivers/nvmem/ |
H A D | qcom-spmi-sdam.c | 64 size_t bytes) 70 if (!sdam_is_valid(sdam, offset, bytes)) { 72 offset, bytes); 76 rc = regmap_bulk_read(sdam->regmap, sdam->base + offset, val, bytes); 79 offset, bytes, rc); 85 size_t bytes) 91 if (!sdam_is_valid(sdam, offset, bytes)) { 93 offset, bytes); 97 if (sdam_is_ro(offset, bytes)) { 99 offset, bytes); 63 sdam_read(void *priv, unsigned int offset, void *val, size_t bytes) argument 84 sdam_write(void *priv, unsigned int offset, void *val, size_t bytes) argument [all...] |
H A D | meson-efuse.c | 18 void *val, size_t bytes) 22 return meson_sm_call_read(fw, (u8 *)val, bytes, SM_EFUSE_READ, offset, 23 bytes, 0, 0, 0); 27 void *val, size_t bytes) 31 return meson_sm_call_write(fw, (u8 *)val, bytes, SM_EFUSE_WRITE, offset, 32 bytes, 0, 0, 0); 17 meson_efuse_read(void *context, unsigned int offset, void *val, size_t bytes) argument 26 meson_efuse_write(void *context, unsigned int offset, void *val, size_t bytes) argument
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | hash-utils.h | 29 u64 bytes = (u64) get_unaligned_be16(chapter_bits) << 32; local 31 bytes |= get_unaligned_be32(chapter_bits + 2); 32 return bytes;
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_cmd.c | 187 static bool vmw_fifo_is_full(struct vmw_private *dev_priv, uint32_t bytes) argument 194 return ((max - next_cmd) + (stop - min) <= bytes); 198 uint32_t bytes, bool interruptible, 211 if (!vmw_fifo_is_full(dev_priv, bytes)) 231 uint32_t bytes, bool interruptible, 236 if (likely(!vmw_fifo_is_full(dev_priv, bytes))) 241 return vmw_fifo_wait_noirq(dev_priv, bytes, 250 !vmw_fifo_is_full(dev_priv, bytes), timeout); 254 !vmw_fifo_is_full(dev_priv, bytes), timeout); 268 * Reserve @bytes numbe 197 vmw_fifo_wait_noirq(struct vmw_private *dev_priv, uint32_t bytes, bool interruptible, unsigned long timeout) argument 230 vmw_fifo_wait(struct vmw_private *dev_priv, uint32_t bytes, bool interruptible, unsigned long timeout) argument 277 vmw_local_fifo_reserve(struct vmw_private *dev_priv, uint32_t bytes) argument 365 vmw_cmd_ctx_reserve(struct vmw_private *dev_priv, uint32_t bytes, int ctx_id) argument 385 vmw_fifo_res_copy(struct vmw_fifo_state *fifo_state, struct vmw_private *vmw, uint32_t next_cmd, uint32_t max, uint32_t min, uint32_t bytes) argument 407 vmw_fifo_slow_copy(struct vmw_fifo_state *fifo_state, struct vmw_private *vmw, uint32_t next_cmd, uint32_t max, uint32_t min, uint32_t bytes) argument 427 vmw_local_fifo_commit(struct vmw_private *dev_priv, uint32_t bytes) argument 472 vmw_cmd_commit(struct vmw_private *dev_priv, uint32_t bytes) argument 487 vmw_cmd_commit_flush(struct vmw_private *dev_priv, uint32_t bytes) argument 517 uint32_t bytes = sizeof(u32) + sizeof(*cmd_fence); local [all...] |
/linux-master/lib/crypto/ |
H A D | poly1305.c | 33 unsigned int bytes; local 36 bytes = min(nbytes, POLY1305_BLOCK_SIZE - desc->buflen); 37 memcpy(desc->buf + desc->buflen, src, bytes); 38 src += bytes; 39 nbytes -= bytes; 40 desc->buflen += bytes;
|
/linux-master/lib/raid6/ |
H A D | recov_neon.c | 23 static void raid6_2data_recov_neon(int disks, size_t bytes, int faila, argument 45 raid6_call.gen_syndrome(disks, bytes, ptrs); 59 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); 63 static void raid6_datap_recov_neon(int disks, size_t bytes, int faila, argument 80 raid6_call.gen_syndrome(disks, bytes, ptrs); 90 __raid6_datap_recov_neon(bytes, p, q, dq, qmul);
|
H A D | recov_s390xc.c | 23 static void raid6_2data_recov_s390xc(int disks, size_t bytes, int faila, argument 44 raid6_call.gen_syndrome(disks, bytes, ptrs); 57 while (bytes) { 67 bytes -= 256; 72 static void raid6_datap_recov_s390xc(int disks, size_t bytes, int faila, argument 88 raid6_call.gen_syndrome(disks, bytes, ptrs); 98 while (bytes) { 106 bytes -= 256;
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | netcnt.c | 19 unsigned long bytes; local 61 bytes = netcnt.bytes; 66 bytes += percpu_netcnt[cpu].bytes; 72 /* Let's check that bytes counter matches the number of packets 75 ASSERT_GE(bytes, packets * 104, "bytes");
|
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | mmio.h | 90 void *p_data, unsigned int bytes); 92 void *p_data, unsigned int bytes); 95 void *p_data, unsigned int bytes); 97 void *p_data, unsigned int bytes); 103 void *pdata, unsigned int bytes, bool is_read); 106 void *p_data, unsigned int bytes);
|
H A D | page_track.h | 35 u64 gpa, void *data, int bytes); 57 void *data, unsigned int bytes);
|
/linux-master/sound/i2c/other/ |
H A D | pt2258.c | 34 unsigned char bytes[2]; local 38 bytes[0] = PT2258_CMD_RESET; 40 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) 46 bytes[0] = PT2258_CMD_MUTE; 48 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) 55 bytes[0] = 0xd0; 56 bytes[1] = 0xe0; 58 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 2) != 2) 97 unsigned char bytes[2]; local 108 bytes[ 146 unsigned char bytes[2]; local [all...] |