/linux-master/fs/gfs2/ |
H A D | rgrp.c | 380 u32 chunk_size; local 402 chunk_size = ((ptr == NULL) ? bytes : (ptr - start)); 403 chunk_size *= GFS2_NBBY; 404 BUG_ON(len < chunk_size); 405 len -= chunk_size; 407 if (gfs2_rbm_from_block(&rbm, block + chunk_size)) {
|
/linux-master/fs/btrfs/ |
H A D | volumes.c | 5256 u64 chunk_size; member in struct:alloc_chunk_ctl 5269 ctl->max_chunk_size = READ_ONCE(space_info->chunk_size); 5461 ctl->chunk_size = ctl->stripe_size * data_stripes; 5493 ctl->chunk_size = ctl->stripe_size * data_stripes; 5636 map->chunk_len = ctl->chunk_size; 5653 trace_btrfs_chunk_alloc(info, map, start, ctl->chunk_size); 5661 block_group = btrfs_make_block_group(trans, type, start, ctl->chunk_size);
|
H A D | sysfs.c | 747 return sysfs_emit(buf, "%llu\n", READ_ONCE(sinfo->chunk_size)); 897 BTRFS_ATTR_RW(space_info, chunk_size, btrfs_chunk_size_show, btrfs_chunk_size_store); 950 BTRFS_ATTR_PTR(space_info, chunk_size),
|
H A D | block-group.c | 2649 u64 chunk_offset, u64 chunk_size) 2658 map = btrfs_get_chunk_map(fs_info, chunk_offset, chunk_size); 2648 insert_dev_extents(struct btrfs_trans_handle *trans, u64 chunk_offset, u64 chunk_size) argument
|
/linux-master/drivers/ufs/core/ |
H A D | ufshcd.c | 2232 (!m->chunk_size || m->chunk_size == lrbp->cmd->sdb.length) &&
|
/linux-master/drivers/scsi/smartpqi/ |
H A D | smartpqi_init.c | 8810 static int pqi_ofa_alloc_mem(struct pqi_ctrl_info *ctrl_info, u32 total_size, u32 chunk_size) argument 8821 sg_count = DIV_ROUND_UP(total_size, chunk_size); 8833 dma_alloc_coherent(dev, chunk_size, &dma_handle, GFP_KERNEL); 8838 put_unaligned_le32(chunk_size, &mem_descriptor->length); 8843 put_unaligned_le32(sg_count * chunk_size, &ofap->bytes_allocated); 8850 dma_free_coherent(dev, chunk_size, 8863 u32 chunk_size; local 8873 for (chunk_size = total_size; chunk_size >= min_chunk_size;) { 8874 if (pqi_ofa_alloc_mem(ctrl_info, total_size, chunk_size) [all...] |
/linux-master/drivers/scsi/megaraid/ |
H A D | megaraid_sas_fusion.c | 795 u32 chunk_size, array_size, offset; local 798 chunk_size = fusion->reply_alloc_sz * RDPQ_MAX_INDEX_IN_ONE_CHUNK; 816 chunk_size, 16, 0); 820 chunk_size, 821 roundup_pow_of_two(chunk_size), 863 chunk_size)) {
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 542 u32 xsk_chunk_size = rq->xsk_pool ? rq->xsk_pool->chunk_size : 0; 4422 max_mtu_frame = MLX5E_HW2SW_MTU(new_params, xsk.chunk_size - hr);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | setup.c | 32 * and xsk->chunk_size is limited to 65535 bytes. 34 if ((size_t)xsk->chunk_size > PAGE_SIZE || xsk->chunk_size < MLX5E_MIN_XSK_CHUNK_SIZE) { 35 mlx5_core_err(mdev, "XSK chunk size %u out of bounds [%u, %lu]\n", xsk->chunk_size,
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.c | 22 u8 req_page_shift = xsk ? order_base_2(xsk->chunk_size) : PAGE_SHIFT; 49 oversized = xsk->chunk_size < (1 << page_shift); 50 WARN_ON_ONCE(xsk->chunk_size > (1 << page_shift)); 66 if (xsk->chunk_size % 3 == 0 && is_power_of_2(xsk->chunk_size / 3)) 314 if (xsk && mlx5e_rx_get_linear_sz_xsk(params, xsk) > xsk->chunk_size) 569 1 << params->log_rq_mtu_frames, xsk->chunk_size); 1134 .chunk_size = 1 << frame_shift, 1148 xsk.chunk_size -= 1; 1153 xsk.chunk_size [all...] |
H A D | params.h | 11 u16 chunk_size; member in struct:mlx5e_xsk_param
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_adminq_cmd.h | 1010 u8 chunk_size; member in struct:ice_aqc_layer_props
|
/linux-master/drivers/md/ |
H A D | md.c | 1386 mddev->chunk_sectors = sb->chunk_size >> 9; 1562 sb->chunk_size = mddev->chunk_sectors << 9; 4339 __ATTR(chunk_size, S_IRUGO|S_IWUSR, chunk_size_show, chunk_size_store); 5285 /* Must be a multiple of chunk_size */ 6787 info.chunk_size = mddev->chunk_sectors << 9; 7337 mddev->chunk_sectors = info->chunk_size >> 9; 7462 * The version, ctime,level,size,raid_disks,not_persistent, layout,chunk_size 7484 mddev->chunk_sectors != info->chunk_size >> 9 ||
|
H A D | dm-snap.c | 538 if (l->store->chunk_size < s->store->chunk_size) 858 unsigned int chunk_size = rounddown_pow_of_two(UINT_MAX); local 862 chunk_size = min_not_zero(chunk_size, 863 snap->store->chunk_size); 865 return (uint32_t) chunk_size; 1017 sector_t sector, unsigned int chunk_size); 1082 io_size = linear_chunks * s->store->chunk_size; 1378 s->store->chunk_size [all...] |
/linux-master/drivers/s390/cio/ |
H A D | css.c | 1092 size_t chunk_size = chunk->end_addr - chunk->start_addr + 1; local 1094 dma_free_coherent((struct device *) data, chunk_size, 1121 size_t chunk_size; local 1128 chunk_size = round_up(size, PAGE_SIZE); 1129 addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP); 1132 gen_pool_add_virt(gp_dma, (unsigned long)addr, dma_addr, chunk_size, -1);
|
/linux-master/drivers/nvme/host/ |
H A D | pci.c | 1940 u32 chunk_size) 1949 tmp = (preferred + chunk_size - 1); 1950 do_div(tmp, chunk_size); 1968 len = min_t(u64, chunk_size, preferred - size); 2011 u64 chunk_size; local 2014 for (chunk_size = min_chunk; chunk_size >= hmminds; chunk_size /= 2) { 2015 if (!__nvme_alloc_host_mem(dev, preferred, chunk_size)) { 1939 __nvme_alloc_host_mem(struct nvme_dev *dev, u64 preferred, u32 chunk_size) argument
|
/linux-master/lib/ |
H A D | scatterlist.c | 514 unsigned int j, chunk_size; local 537 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; 539 min_t(unsigned long, size, chunk_size), offset); 541 size -= chunk_size;
|
/linux-master/fs/nilfs2/ |
H A D | dir.c | 113 unsigned int chunk_size = nilfs_chunk_size(dir); local 121 if (limit & (chunk_size - 1)) 136 if (((offs + rec_len - 1) ^ offs) & ~(chunk_size-1)) 427 unsigned int chunk_size = nilfs_chunk_size(dir); local 456 rec_len = chunk_size; 457 de->rec_len = nilfs_rec_len_to_disk(chunk_size); 562 unsigned int chunk_size = nilfs_chunk_size(inode); local 570 err = nilfs_prepare_chunk(folio, 0, chunk_size); 576 memset(kaddr, 0, chunk_size); 586 de->rec_len = nilfs_rec_len_to_disk(chunk_size [all...] |
/linux-master/net/xdp/ |
H A D | xsk.c | 639 ts = pool->unaligned ? len : pool->chunk_size; 1333 __u32 chunk_size; member in struct:xdp_umem_reg_v1 1340 __u32 chunk_size; member in struct:xdp_umem_reg_v2
|
/linux-master/tools/testing/selftests/net/ |
H A D | tls.c | 434 uint16_t chunk_size, 447 EXPECT_GE(chunk_size, 1); 448 test_payload_size = chunk_size + extra_payload_size; 456 ret = sendfile(self->fd, fd, &offset, chunk_size); 432 chunked_sendfile(struct __test_metadata *_metadata, struct _test_data_tls *self, uint16_t chunk_size, uint16_t extra_payload_size) argument
|
/linux-master/tools/testing/selftests/mm/ |
H A D | uffd-unit-tests.c | 1113 uffd_move_test_common(uffd_test_args_t *targs, unsigned long chunk_size, argument 1138 step_size = chunk_size / page_size; 1141 if (chunk_size > page_size) { 1142 char *aligned_src = ALIGN_UP(area_src, chunk_size); 1143 char *aligned_dst = ALIGN_UP(area_dst, chunk_size);
|
/linux-master/include/net/ |
H A D | xdp_sock.h | 27 u32 chunk_size; member in struct:xdp_umem
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | verbs.c | 2820 stag_info->chunk_size = 1; 2823 stag_info->chunk_size = 3; 3560 stag_info.chunk_size = 1;
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hem.c | 208 u32 chunk_size; local 224 chunk_size = table->type < HEM_TYPE_MTT ? mhop->buf_chunk_size : 226 table_idx = *obj / (chunk_size / table->obj_size);
|
/linux-master/arch/x86/xen/ |
H A D | setup.c | 688 phys_addr_t mem_end, addr, size, chunk_size; local 773 chunk_size = size; 781 chunk_size = min(size, mem_end - addr); 783 chunk_size = min(size, PFN_PHYS(extra_pages)); 785 n_pfns = PFN_DOWN(addr + chunk_size) - pfn_s; 794 xen_align_and_add_e820_region(addr, chunk_size, type); 796 addr += chunk_size; 797 size -= chunk_size;
|