/linux-master/net/ncsi/ |
H A D | ncsi-cmd.c | 80 cmd->hw_arbitration = nca->bytes[0]; 92 cmd->ald = nca->bytes[0]; 115 cmd->mc_id = nca->bytes[0]; 142 cmd->index = nca->bytes[6]; 143 cmd->enable = nca->bytes[7]; 155 cmd->mode = nca->bytes[3]; 169 cmd->mac[i] = nca->bytes[i]; 170 cmd->index = nca->bytes[6]; 171 cmd->at_e = nca->bytes[7]; 207 cmd->mode = nca->bytes[ [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | bpf_tcp_ca.c | 39 ssize_t nr_sent = 0, bytes = 0; local 55 while (bytes < total_bytes && !READ_ONCE(stop)) { 57 MIN(total_bytes - bytes, sizeof(batch)), 0); 64 bytes += nr_sent; 67 ASSERT_EQ(bytes, total_bytes, "send"); 82 ssize_t nr_recv = 0, bytes = 0; local 148 while (bytes < total_bytes && !READ_ONCE(stop)) { 150 MIN(total_bytes - bytes, sizeof(batch)), 0); 155 bytes += nr_recv; 158 ASSERT_EQ(bytes, total_byte [all...] |
/linux-master/drivers/spi/ |
H A D | spi-omap-uwire.c | 206 unsigned bytes; local 221 /* write one or two bytes at a time */ 228 bytes = 2; 231 bytes = 1; 247 len -= bytes; 256 status += bytes; 266 /* read one or two bytes at a time */ 269 bytes = 2; 271 bytes = 1; 276 len -= bytes; [all...] |
/linux-master/fs/ntfs3/ |
H A D | fsntfs.c | 131 bool ntfs_fix_pre_write(struct NTFS_RECORD_HEADER *rhdr, size_t bytes) argument 139 fn * SECTOR_SIZE > bytes) { 168 int ntfs_fix_post_read(struct NTFS_RECORD_HEADER *rhdr, size_t bytes, argument 176 fn = simple ? ((bytes >> SECTOR_SHIFT) + 1) : 181 fn * SECTOR_SIZE > bytes) { 850 u32 blocksize, bytes; local 861 bytes = sbi->mft.recs_mirr << sbi->record_bits; 865 for (; bytes >= blocksize; bytes -= blocksize) { 999 static inline __le32 security_hash(const void *sd, size_t bytes) argument 1034 ntfs_sb_read(struct super_block *sb, u64 lbo, size_t bytes, void *buffer) argument 1062 ntfs_sb_write(struct super_block *sb, u64 lbo, size_t bytes, const void *buf, int wait) argument 1127 ntfs_sb_write_run(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, const void *buf, size_t bytes, int sync) argument 1191 ntfs_read_run_nb(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb) argument 1309 ntfs_read_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, struct NTFS_RECORD_HEADER *rhdr, u32 bytes, struct ntfs_buffers *nb) argument 1320 ntfs_get_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, u32 bytes, struct ntfs_buffers *nb) argument 1421 u32 bytes = nb->bytes; local 1497 ntfs_bio_pages(struct ntfs_sb_info *sbi, const struct runs_tree *run, struct page **pages, u32 nr_pages, u64 vbo, u32 bytes, enum req_op op) argument 1666 ntfs_vbo_to_lbo(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, u64 *lbo, u64 *bytes) argument [all...] |
/linux-master/fs/btrfs/ |
H A D | space-info.c | 40 * determine the size of the block reserves, and then use the actual bytes 68 * create a reserve_ticket with ->bytes set to our reservation, add it to 72 * wait on ticket->wait for ->bytes to be reduced to 0, or ->error to be set 87 * Check if ->bytes == 0, if it does we got our reservation and we can carry 111 * isize or bytes. We keep these delayed items to coalesce these operations 412 struct btrfs_space_info *space_info, u64 bytes, 425 if (used + bytes < space_info->total_bytes + avail) 435 ASSERT(space_info->reclaim_size >= ticket->bytes); 436 space_info->reclaim_size -= ticket->bytes; 461 if ((used + ticket->bytes < 411 btrfs_can_overcommit(struct btrfs_fs_info *fs_info, struct btrfs_space_info *space_info, u64 bytes, enum btrfs_reserve_flush_enum flush) argument 535 btrfs_dump_space_info(struct btrfs_fs_info *fs_info, struct btrfs_space_info *info, u64 bytes, int dump_block_groups) argument 580 u64 bytes; local 1803 btrfs_reserve_data_bytes(struct btrfs_fs_info *fs_info, u64 bytes, enum btrfs_reserve_flush_enum flush) argument [all...] |
/linux-master/drivers/net/ethernet/google/gve/ |
H A D | gve_tx.c | 64 size_t bytes) 66 return (fifo->head + bytes < fifo->size) ? 0 : fifo->size - fifo->head; 69 static bool gve_tx_fifo_can_alloc(struct gve_tx_fifo *fifo, size_t bytes) argument 71 return (atomic_read(&fifo->available) <= bytes) ? false : true; 76 * @bytes: Allocation size 84 static int gve_tx_alloc_fifo(struct gve_tx_fifo *fifo, size_t bytes, argument 91 if (!bytes) 100 WARN(!gve_tx_fifo_can_alloc(fifo, bytes), 106 iov[0].iov_len = bytes; 107 fifo->head += bytes; 63 gve_tx_fifo_pad_alloc_one_frag(struct gve_tx_fifo *fifo, size_t bytes) argument 139 gve_tx_free_fifo(struct gve_tx_fifo *fifo, size_t bytes) argument 162 u64 pkts = 0, bytes = 0; local 218 size_t bytes; local 259 size_t bytes; local 401 int bytes; local 878 u64 pkts = 0, bytes = 0; local [all...] |
/linux-master/fs/ocfs2/ |
H A D | ocfs2.h | 149 /* Storing max wait in usecs saves 24 bytes per inode */ 744 u64 bytes) 749 bytes += OCFS2_SB(sb)->s_clustersize - 1; 751 clusters = (unsigned int)(bytes >> cl_bits); 757 u64 bytes) 762 clusters = (unsigned int)(bytes >> cl_bits); 767 u64 bytes) 769 bytes += sb->s_blocksize - 1; 770 return bytes >> sb->s_blocksize_bits; 790 u64 bytes) 743 ocfs2_clusters_for_bytes(struct super_block *sb, u64 bytes) argument 756 ocfs2_bytes_to_clusters(struct super_block *sb, u64 bytes) argument 766 ocfs2_blocks_for_bytes(struct super_block *sb, u64 bytes) argument 789 ocfs2_align_bytes_to_clusters(struct super_block *sb, u64 bytes) argument 799 ocfs2_align_bytes_to_blocks(struct super_block *sb, u64 bytes) argument 808 ocfs2_align_bytes_to_sectors(u64 bytes) argument [all...] |
/linux-master/drivers/staging/media/atomisp/pci/runtime/frame/src/ |
H A D | frame.c | 576 height = frame->frame_info.res.height, stride, bytes; local 579 bytes = stride * height; 580 frame->data_bytes = 3 * bytes; 583 width, stride, height, 1 * bytes); 585 width, stride, height, 2 * bytes); 592 height = frame->frame_info.res.height / 2, bytes, stride; local 595 bytes = stride * height; 597 frame->data_bytes = 6 * bytes; 599 width, stride, height, 0 * bytes); 601 width, stride, height, 1 * bytes); [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | fs_counters.c | 50 u64 bytes; member in struct:mlx5_fc_cache 59 /* last{packets,bytes} members are used when calculating the delta since 102 * query/dump - packets and bytes might be inconsistent (since update is not 167 u64 bytes = MLX5_GET64(traffic_counter, stats, octets); local 173 cache->bytes = bytes; 365 counter->lastbytes = counter->cache.bytes; 481 u64 *packets, u64 *bytes) 483 return mlx5_cmd_fc_query(dev, counter->id, packets, bytes); 493 u64 *bytes, u6 480 mlx5_fc_query(struct mlx5_core_dev *dev, struct mlx5_fc *counter, u64 *packets, u64 *bytes) argument 492 mlx5_fc_query_cached(struct mlx5_fc *counter, u64 *bytes, u64 *packets, u64 *lastuse) argument 507 mlx5_fc_query_cached_raw(struct mlx5_fc *counter, u64 *bytes, u64 *packets, u64 *lastuse) argument [all...] |
/linux-master/drivers/nvmem/ |
H A D | rockchip-otp.c | 148 void *val, size_t bytes) 168 while (bytes--) { 188 void *val, size_t bytes) 197 addr_end = round_up(offset + bytes, RK3588_NBYTES) / RK3588_NBYTES; 225 memcpy(val, buf + offset % RK3588_NBYTES, bytes); 234 void *val, size_t bytes) 248 ret = otp->data->reg_read(context, offset, val, bytes); 147 px30_otp_read(void *context, unsigned int offset, void *val, size_t bytes) argument 187 rk3588_otp_read(void *context, unsigned int offset, void *val, size_t bytes) argument 233 rockchip_otp_read(void *context, unsigned int offset, void *val, size_t bytes) argument
|
H A D | qfprom.c | 250 * @bytes: The number of bytes to write. 257 size_t bytes) 261 int words = bytes / 4; 269 reg, bytes); 276 if (bytes % 4) { 278 "%zu is not an integral number of words\n", bytes); 321 unsigned int reg, void *_val, size_t bytes) 325 int i = 0, words = bytes; 256 qfprom_reg_write(void *context, unsigned int reg, void *_val, size_t bytes) argument 320 qfprom_reg_read(void *context, unsigned int reg, void *_val, size_t bytes) argument
|
H A D | imx-ocotp.c | 158 void *val, size_t bytes) 167 num_bytes = round_up((offset % 4) + bytes, 4); 212 memcpy(val, &p[index], bytes); 224 unsigned int offset, void *data, size_t bytes) 231 for (i = 0; i < bytes / 2; i++) 232 swap(buf[i], buf[bytes - i - 1]); 311 size_t bytes) 322 if ((bytes != priv->config->word_size) || 474 return ret < 0 ? ret : bytes; 157 imx_ocotp_read(void *context, unsigned int offset, void *val, size_t bytes) argument 223 imx_ocotp_cell_pp(void *context, const char *id, int index, unsigned int offset, void *data, size_t bytes) argument 310 imx_ocotp_write(void *context, unsigned int offset, void *val, size_t bytes) argument
|
H A D | core.c | 31 int bytes; member in struct:nvmem_cell_entry 59 void *val, size_t bytes) 62 return nvmem->reg_read(nvmem->priv, offset, val, bytes); 68 void *val, size_t bytes) 74 ret = nvmem->reg_write(nvmem->priv, offset, val, bytes); 84 size_t bytes, int write) 87 unsigned int end = offset + bytes; 147 void *val, size_t bytes) 150 return __nvmem_reg_read(nvmem, offset, val, bytes); 152 return nvmem_access_with_keepouts(nvmem, offset, val, bytes, fals 58 __nvmem_reg_read(struct nvmem_device *nvmem, unsigned int offset, void *val, size_t bytes) argument 67 __nvmem_reg_write(struct nvmem_device *nvmem, unsigned int offset, void *val, size_t bytes) argument 82 nvmem_access_with_keepouts(struct nvmem_device *nvmem, unsigned int offset, void *val, size_t bytes, int write) argument 146 nvmem_reg_read(struct nvmem_device *nvmem, unsigned int offset, void *val, size_t bytes) argument 155 nvmem_reg_write(struct nvmem_device *nvmem, unsigned int offset, void *val, size_t bytes) argument 2051 nvmem_device_read(struct nvmem_device *nvmem, unsigned int offset, size_t bytes, void *buf) argument 2079 nvmem_device_write(struct nvmem_device *nvmem, unsigned int offset, size_t bytes, void *buf) argument [all...] |
/linux-master/arch/x86/lib/ |
H A D | checksum_32.S | 58 jz 10f # Jump if alignment is boundary of 2 bytes. 70 subl $2, %ecx # Alignment uses up two bytes. 71 jae 1f # Jump if we had at least two bytes. 187 addw (%esi), %ax # csumming 2 bytes, 2-aligned 233 # Handle the last 1-3 bytes without jumping 284 subl $2, %ecx # Alignment uses up two bytes. 285 jae 1f # Jump if we had at least two bytes.
|
/linux-master/lib/ |
H A D | test_meminit.c | 28 /* Calculate the number of uninitialized bytes in the buffer. */ 40 /* Fill a buffer with garbage, skipping |skip| first bytes. */ 171 /* Initialize the first 4 bytes of the object. */ 179 * If the cache has a test constructor, the first 4 bytes of the object must 190 int bytes; local 193 bytes = count_nonzero_bytes(buf, size); 196 return bytes; 201 if (bytes) 351 int num, bytes; local 360 bytes [all...] |
H A D | string.c | 133 * read_word_at_a_time() below may read uninitialized bytes after the 282 * @count: The maximum number of bytes to compare 536 * store, not the number of bytes. 558 * store, not the number of bytes. 580 * store, not the number of bytes. 800 static void *check_bytes8(const u8 *start, u8 value, unsigned int bytes) argument 802 while (bytes) { 806 bytes--; 815 * @bytes: The size of the area. 820 void *memchr_inv(const void *start, int c, size_t bytes) argument [all...] |
/linux-master/fs/iomap/ |
H A D | buffered-io.c | 874 /* Returns the number of bytes copied. May be 0. Cannot be an errno. */ 922 size_t bytes; /* Bytes to write to folio */ local 925 bytes = iov_iter_count(i); 928 bytes = min(chunk - offset, bytes); 934 if (bytes > length) 935 bytes = length; 947 if (unlikely(fault_in_iov_iter_readable(i, bytes) == bytes)) { 952 status = iomap_write_begin(iter, pos, bytes, 1324 size_t bytes = min_t(u64, SIZE_MAX, length); local 1385 size_t bytes = min_t(u64, SIZE_MAX, length); local [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_qp_tx.c | 53 u32 bytes = sge->length; local 55 if (bytes > MAX_HDR_INLINE || wqe->sqe.num_sge != 1) 58 if (!bytes) 62 memcpy(paddr, &wqe->sqe.sge[1], bytes); 68 memcpy(paddr, ib_virt_dma_to_ptr(sge->laddr), bytes); 71 bytes)) 85 if (likely(PAGE_SIZE - off >= bytes)) { 86 memcpy(paddr, buffer + off, bytes); 88 unsigned long part = bytes - (PAGE_SIZE - off); 98 memcpy(paddr + part, buffer, bytes 328 size_t bytes = min_t(size_t, PAGE_SIZE - offset, size); local [all...] |
/linux-master/net/vmw_vsock/ |
H A D | virtio_transport_common.c | 430 /* Return number of bytes, if any data has been sent. */ 518 size_t bytes; local 520 bytes = len - total; 521 if (bytes > skb->len) 522 bytes = skb->len; 530 &msg->msg_iter, bytes); 534 total += bytes; 558 size_t bytes, total = 0; local 576 bytes = min_t(size_t, len - total, 586 &msg->msg_iter, bytes); 655 size_t bytes; local 825 s64 bytes; local 851 s64 bytes; local 863 s64 bytes; local [all...] |
/linux-master/arch/x86/boot/ |
H A D | header.S | 234 # Explicitly enter this as bytes, or the assembler 280 ramdisk_size: .long 0 # its size in bytes 409 # ... resulting in +18 bytes overhead of uncompressed data. 417 # 1 block occurs every 32K -1 bytes or when there 50% compression 421 # 32 bits length in bytes. 445 # of 5 bytes per 32767 bytes. 449 # 32764 bytes and then all of the rest of the bytes representing the very 453 # to be safe. To avoid problems at the block level allocating 5 extra bytes [all...] |
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmw_surface_cache.h | 105 * Return the number of bytes of buffer space required to store one image of a 164 * vmw_surface_get_serialized_size_extended - Returns the number of bytes 185 * vmw_surface_get_pixel_offset - Compute the offset (in bytes) to a pixel 222 u32 bytes; local 230 bytes = vmw_surface_get_image_buffer_size(desc, &mipSize, 0); 231 mipChainBytes += bytes; 233 mipChainBytesToLevel += bytes; 304 * @bytes: Bytes required in the backing store of this mipmap level. 310 size_t bytes; member in struct:vmw_surface_mip 396 mip->bytes [all...] |
/linux-master/drivers/dma/hsu/ |
H A D | hsu.c | 312 size_t bytes = 0; local 316 bytes += desc->sg[i].len; 320 bytes += hsu_chan_readl(hsuc, HSU_CH_DxTSR(i)); 323 return bytes; 332 size_t bytes; local 342 bytes = hsu_dma_active_desc_size(hsuc); 343 dma_set_residue(state, bytes); 346 bytes = to_hsu_dma_desc(vdesc)->length; 347 dma_set_residue(state, bytes);
|
/linux-master/net/ceph/ |
H A D | messenger.c | 751 size_t bytes) 756 BUG_ON(bytes > cursor->resid); 757 BUG_ON(bytes > bio_iter_len(it->bio, it->iter)); 758 cursor->resid -= bytes; 759 bio_advance_iter(it->bio, &it->iter, bytes); 764 if (!bytes || (it->iter.bi_size && it->iter.bi_bvec_done && 766 return false; /* more bytes to process in this segment */ 806 size_t bytes) 811 BUG_ON(bytes > cursor->resid); 812 BUG_ON(bytes > bvec_iter_le 750 ceph_msg_data_bio_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument 805 ceph_msg_data_bvecs_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument 867 ceph_msg_data_pages_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument 940 ceph_msg_data_pagelist_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument 1012 ceph_msg_data_iter_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument 1126 ceph_msg_data_advance(struct ceph_msg_data_cursor *cursor, size_t bytes) argument [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grukdump.c | 115 int try, cch_locked, cbrcnt = 0, dsrcnt = 0, bytes = 0, ret = 0; local 137 bytes = sizeof(hdr) + GRU_CACHE_LINE_BYTES; 151 bytes += (3 * cbrcnt + dsrcnt) * GRU_CACHE_LINE_BYTES; 152 if (bytes > ubufend - ubuf) 172 return bytes;
|
/linux-master/samples/bpf/ |
H A D | cookie_uid_helper_example.c | 61 uint64_t bytes; member in struct:stats 129 -32 + (__s16)offsetof(struct stats, bytes)), 155 offsetof(struct stats, bytes)), 212 curEntry.packets, curEntry.bytes); 270 dataEntry.packets, dataEntry.bytes);
|