/linux-master/drivers/accessibility/speakup/ |
H A D | devsynth.c | 19 size_t bytes; local 26 bytes = min(count, sizeof(buf)); 27 if (copy_from_user(buf, ptr, bytes)) 29 count -= bytes; 30 ptr += bytes; 32 synth_write(buf, bytes); 44 size_t bytes; local 56 bytes = min(count, sizeof(buf)); 57 if (copy_from_user(buf, ptr, bytes)) 61 for (in = 0, out = 0; in < bytes; i [all...] |
/linux-master/net/rds/ |
H A D | page.c | 52 * @bytes: the number of bytes needed. 58 * If @bytes is at least a full page then this just returns a page from 61 * If @bytes is a partial page then this stores the unused region of the 68 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, argument 79 if (bytes >= PAGE_SIZE) { 95 if (rem->r_page && bytes > (PAGE_SIZE - rem->r_offset)) { 102 if (rem->r_page && bytes <= (PAGE_SIZE - rem->r_offset)) { 103 sg_set_page(scat, rem->r_page, bytes, rem->r_offset); 109 rem->r_offset += ALIGN(bytes, [all...] |
/linux-master/include/uapi/linux/netfilter/ |
H A D | xt_sctp.h | 39 #define bytes(type) (sizeof(type) * 8) macro 43 (chunkmap)[type / bytes(__u32)] |= \ 44 1u << (type % bytes(__u32)); \ 49 (chunkmap)[type / bytes(__u32)] &= \ 50 ~(1u << (type % bytes(__u32))); \ 55 ((chunkmap)[type / bytes (__u32)] & \ 56 (1u << (type % bytes (__u32)))) ? 1: 0; \
|
/linux-master/drivers/char/hw_random/ |
H A D | arm_smccc_trng.c | 37 size_t bytes) 41 if (bytes == 0) 44 chunk = min(bytes, sizeof(long)); 47 if (copied >= bytes) 50 chunk = min((bytes - copied), sizeof(long)); 53 if (copied >= bytes) 56 chunk = min((bytes - copied), sizeof(long)); 36 copy_from_registers(char *buf, struct arm_smccc_res *res, size_t bytes) argument
|
/linux-master/net/netfilter/ |
H A D | xt_connbytes.c | 16 MODULE_DESCRIPTION("Xtables: Number of packets/bytes per connection matching"); 27 u_int64_t bytes = 0; local 59 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 62 what = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); 65 what = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 66 what += atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); 73 bytes = atomic64_read(&counters[IP_CT_DIR_ORIGINAL].bytes); 77 bytes = atomic64_read(&counters[IP_CT_DIR_REPLY].bytes); [all...] |
/linux-master/fs/btrfs/ |
H A D | space-info.h | 99 u64 total_bytes; /* total bytes in the space, 101 u64 bytes_used; /* total bytes used, 103 u64 bytes_pinned; /* total bytes pinned, will be freed when the 105 u64 bytes_reserved; /* total bytes the allocator has reserved for 107 u64 bytes_may_use; /* number of bytes that may be used for 109 u64 bytes_readonly; /* total bytes that are read only */ 110 u64 bytes_zone_unusable; /* total bytes that are unusable until 116 /* Chunk size in bytes */ 138 u64 disk_used; /* total bytes used on disk */ 139 u64 disk_total; /* total bytes o 171 u64 bytes; member in struct:reserve_ticket [all...] |
/linux-master/arch/riscv/crypto/ |
H A D | ghash-riscv64-glue.c | 30 u32 bytes; member in struct:riscv64_ghash_desc_ctx 83 if (dctx->bytes) { 84 if (dctx->bytes + srclen < GHASH_BLOCK_SIZE) { 85 memcpy(dctx->buffer + dctx->bytes, src, srclen); 86 dctx->bytes += srclen; 89 memcpy(dctx->buffer + dctx->bytes, src, 90 GHASH_BLOCK_SIZE - dctx->bytes); 93 src += GHASH_BLOCK_SIZE - dctx->bytes; 94 srclen -= GHASH_BLOCK_SIZE - dctx->bytes; 95 dctx->bytes [all...] |
/linux-master/tools/lib/ |
H A D | string.c | 172 static void *check_bytes8(const u8 *start, u8 value, unsigned int bytes) argument 174 while (bytes) { 178 bytes--; 187 * @bytes: The size of the area. 192 void *memchr_inv(const void *start, int c, size_t bytes) argument 198 if (bytes <= 16) 199 return check_bytes8(start, value, bytes); 215 bytes -= prefix; 218 words = bytes / 8; 227 return check_bytes8(start, value, bytes [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | xor.h | 47 xor_arm4regs_2(unsigned long bytes, unsigned long * __restrict p1, argument 50 unsigned int lines = bytes / sizeof(unsigned long) / 4; 68 xor_arm4regs_3(unsigned long bytes, unsigned long * __restrict p1, argument 72 unsigned int lines = bytes / sizeof(unsigned long) / 4; 91 xor_arm4regs_4(unsigned long bytes, unsigned long * __restrict p1, argument 96 unsigned int lines = bytes / sizeof(unsigned long) / 2; 112 xor_arm4regs_5(unsigned long bytes, unsigned long * __restrict p1, argument 118 unsigned int lines = bytes / sizeof(unsigned long) / 2; 156 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, argument 160 xor_arm4regs_2(bytes, p 169 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3) argument 183 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4) argument 198 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, const unsigned long * __restrict p2, const unsigned long * __restrict p3, const unsigned long * __restrict p4, const unsigned long * __restrict p5) argument [all...] |
/linux-master/arch/x86/kvm/mmu/ |
H A D | page_track.h | 30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes); 42 const u8 *new, int bytes) { } 51 const u8 *new, int bytes) 53 __kvm_page_track_write(vcpu->kvm, gpa, new, bytes); 55 kvm_mmu_track_write(vcpu, gpa, new, bytes); 41 __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes) argument 50 kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, int bytes) argument
|
/linux-master/drivers/soc/tegra/fuse/ |
H A D | fuse-tegra30.c | 140 .bytes = 4, 146 .bytes = 4, 152 .bytes = 4, 158 .bytes = 4, 164 .bytes = 4, 170 .bytes = 4, 176 .bytes = 4, 182 .bytes = 4, 188 .bytes = 4, 194 .bytes [all...] |
/linux-master/crypto/ |
H A D | nhpoly1305.c | 88 * 1024 bytes, except possibly the final one which is taken over a multiple of 89 * 16 bytes up to 1024. Also, in the case where data is passed in misaligned 97 unsigned int bytes; local 101 bytes = min_t(unsigned int, srclen, NH_MESSAGE_BYTES); 102 nh_fn(key->nh_key, src, bytes, state->nh_hash); 103 state->nh_remaining = NH_MESSAGE_BYTES - bytes; 111 bytes = min(srclen, state->nh_remaining); 112 nh_fn(&key->nh_key[pos / 4], src, bytes, tmp_hash); 116 state->nh_remaining -= bytes; 120 src += bytes; 161 unsigned int bytes; local [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grukservices.h | 45 * Message queue size is the total number of bytes allocated 52 * bytes size of message queue in bytes 62 void *p, unsigned int bytes, int nasid, int vector, int apicid); 74 * bytes size of message in bytes 82 void *mesg, unsigned int bytes); 141 * bytes number of bytes to copy 148 unsigned int bytes); [all...] |
/linux-master/include/rdma/ |
H A D | iba.h | 46 * aligned, everything larger has a max alignment of 4 bytes. 71 /* FIXME: A set should always set the entire field, meaning we should zero the trailing bytes */ 73 bytes) \ 76 WARN_ON(bytes * 8 > num_bits); \ 77 if (in && bytes) \ 80 _in_ptr, bytes); \ 82 #define IBA_SET_MEM(field, ptr, in, bytes) _IBA_SET_MEM(field, ptr, in, bytes) 94 bytes) \ 97 WARN_ON(bytes * [all...] |
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_cmd.c | 186 static bool vmw_fifo_is_full(struct vmw_private *dev_priv, uint32_t bytes) argument 193 return ((max - next_cmd) + (stop - min) <= bytes); 197 uint32_t bytes, bool interruptible, 210 if (!vmw_fifo_is_full(dev_priv, bytes)) 230 uint32_t bytes, bool interruptible, 235 if (likely(!vmw_fifo_is_full(dev_priv, bytes))) 240 return vmw_fifo_wait_noirq(dev_priv, bytes, 249 !vmw_fifo_is_full(dev_priv, bytes), timeout); 253 !vmw_fifo_is_full(dev_priv, bytes), timeout); 267 * Reserve @bytes numbe 196 vmw_fifo_wait_noirq(struct vmw_private *dev_priv, uint32_t bytes, bool interruptible, unsigned long timeout) argument 229 vmw_fifo_wait(struct vmw_private *dev_priv, uint32_t bytes, bool interruptible, unsigned long timeout) argument 276 vmw_local_fifo_reserve(struct vmw_private *dev_priv, uint32_t bytes) argument 364 vmw_cmd_ctx_reserve(struct vmw_private *dev_priv, uint32_t bytes, int ctx_id) argument 384 vmw_fifo_res_copy(struct vmw_fifo_state *fifo_state, struct vmw_private *vmw, uint32_t next_cmd, uint32_t max, uint32_t min, uint32_t bytes) argument 406 vmw_fifo_slow_copy(struct vmw_fifo_state *fifo_state, struct vmw_private *vmw, uint32_t next_cmd, uint32_t max, uint32_t min, uint32_t bytes) argument 426 vmw_local_fifo_commit(struct vmw_private *dev_priv, uint32_t bytes) argument 471 vmw_cmd_commit(struct vmw_private *dev_priv, uint32_t bytes) argument 486 vmw_cmd_commit_flush(struct vmw_private *dev_priv, uint32_t bytes) argument 516 uint32_t bytes = sizeof(u32) + sizeof(*cmd_fence); local [all...] |
/linux-master/drivers/nvmem/ |
H A D | qcom-spmi-sdam.c | 64 size_t bytes) 70 if (!sdam_is_valid(sdam, offset, bytes)) { 72 offset, bytes); 76 rc = regmap_bulk_read(sdam->regmap, sdam->base + offset, val, bytes); 79 offset, bytes, rc); 85 size_t bytes) 91 if (!sdam_is_valid(sdam, offset, bytes)) { 93 offset, bytes); 97 if (sdam_is_ro(offset, bytes)) { 99 offset, bytes); 63 sdam_read(void *priv, unsigned int offset, void *val, size_t bytes) argument 84 sdam_write(void *priv, unsigned int offset, void *val, size_t bytes) argument [all...] |
H A D | meson-efuse.c | 18 void *val, size_t bytes) 22 return meson_sm_call_read(fw, (u8 *)val, bytes, SM_EFUSE_READ, offset, 23 bytes, 0, 0, 0); 27 void *val, size_t bytes) 31 return meson_sm_call_write(fw, (u8 *)val, bytes, SM_EFUSE_WRITE, offset, 32 bytes, 0, 0, 0); 17 meson_efuse_read(void *context, unsigned int offset, void *val, size_t bytes) argument 26 meson_efuse_write(void *context, unsigned int offset, void *val, size_t bytes) argument
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | hash-utils.h | 29 u64 bytes = (u64) get_unaligned_be16(chapter_bits) << 32; local 31 bytes |= get_unaligned_be32(chapter_bits + 2); 32 return bytes;
|
/linux-master/lib/crypto/ |
H A D | poly1305.c | 33 unsigned int bytes; local 36 bytes = min(nbytes, POLY1305_BLOCK_SIZE - desc->buflen); 37 memcpy(desc->buf + desc->buflen, src, bytes); 38 src += bytes; 39 nbytes -= bytes; 40 desc->buflen += bytes;
|
/linux-master/lib/raid6/ |
H A D | recov_neon.c | 23 static void raid6_2data_recov_neon(int disks, size_t bytes, int faila, argument 45 raid6_call.gen_syndrome(disks, bytes, ptrs); 59 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); 63 static void raid6_datap_recov_neon(int disks, size_t bytes, int faila, argument 80 raid6_call.gen_syndrome(disks, bytes, ptrs); 90 __raid6_datap_recov_neon(bytes, p, q, dq, qmul);
|
H A D | recov_s390xc.c | 23 static void raid6_2data_recov_s390xc(int disks, size_t bytes, int faila, argument 44 raid6_call.gen_syndrome(disks, bytes, ptrs); 57 while (bytes) { 67 bytes -= 256; 72 static void raid6_datap_recov_s390xc(int disks, size_t bytes, int faila, argument 88 raid6_call.gen_syndrome(disks, bytes, ptrs); 98 while (bytes) { 106 bytes -= 256;
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | netcnt.c | 19 unsigned long bytes; local 61 bytes = netcnt.bytes; 66 bytes += percpu_netcnt[cpu].bytes; 72 /* Let's check that bytes counter matches the number of packets 75 ASSERT_GE(bytes, packets * 104, "bytes");
|
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | mmio.h | 90 void *p_data, unsigned int bytes); 92 void *p_data, unsigned int bytes); 95 void *p_data, unsigned int bytes); 97 void *p_data, unsigned int bytes); 103 void *pdata, unsigned int bytes, bool is_read); 106 void *p_data, unsigned int bytes);
|
H A D | page_track.h | 35 u64 gpa, void *data, int bytes); 57 void *data, unsigned int bytes);
|
/linux-master/sound/i2c/other/ |
H A D | pt2258.c | 34 unsigned char bytes[2]; local 38 bytes[0] = PT2258_CMD_RESET; 40 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) 46 bytes[0] = PT2258_CMD_MUTE; 48 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) 55 bytes[0] = 0xd0; 56 bytes[1] = 0xe0; 58 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 2) != 2) 97 unsigned char bytes[2]; local 108 bytes[ 146 unsigned char bytes[2]; local [all...] |