Searched refs:chunk_size (Results 26 - 50 of 151) sorted by last modified time

1234567

/linux-master/fs/gfs2/
H A Drgrp.c380 u32 chunk_size; local
402 chunk_size = ((ptr == NULL) ? bytes : (ptr - start));
403 chunk_size *= GFS2_NBBY;
404 BUG_ON(len < chunk_size);
405 len -= chunk_size;
407 if (gfs2_rbm_from_block(&rbm, block + chunk_size)) {
/linux-master/fs/btrfs/
H A Dvolumes.c5256 u64 chunk_size; member in struct:alloc_chunk_ctl
5269 ctl->max_chunk_size = READ_ONCE(space_info->chunk_size);
5461 ctl->chunk_size = ctl->stripe_size * data_stripes;
5493 ctl->chunk_size = ctl->stripe_size * data_stripes;
5636 map->chunk_len = ctl->chunk_size;
5653 trace_btrfs_chunk_alloc(info, map, start, ctl->chunk_size);
5661 block_group = btrfs_make_block_group(trans, type, start, ctl->chunk_size);
H A Dsysfs.c747 return sysfs_emit(buf, "%llu\n", READ_ONCE(sinfo->chunk_size));
897 BTRFS_ATTR_RW(space_info, chunk_size, btrfs_chunk_size_show, btrfs_chunk_size_store);
950 BTRFS_ATTR_PTR(space_info, chunk_size),
H A Dblock-group.c2649 u64 chunk_offset, u64 chunk_size)
2658 map = btrfs_get_chunk_map(fs_info, chunk_offset, chunk_size);
2648 insert_dev_extents(struct btrfs_trans_handle *trans, u64 chunk_offset, u64 chunk_size) argument
/linux-master/drivers/ufs/core/
H A Dufshcd.c2232 (!m->chunk_size || m->chunk_size == lrbp->cmd->sdb.length) &&
/linux-master/drivers/scsi/smartpqi/
H A Dsmartpqi_init.c8810 static int pqi_ofa_alloc_mem(struct pqi_ctrl_info *ctrl_info, u32 total_size, u32 chunk_size) argument
8821 sg_count = DIV_ROUND_UP(total_size, chunk_size);
8833 dma_alloc_coherent(dev, chunk_size, &dma_handle, GFP_KERNEL);
8838 put_unaligned_le32(chunk_size, &mem_descriptor->length);
8843 put_unaligned_le32(sg_count * chunk_size, &ofap->bytes_allocated);
8850 dma_free_coherent(dev, chunk_size,
8863 u32 chunk_size; local
8873 for (chunk_size = total_size; chunk_size >= min_chunk_size;) {
8874 if (pqi_ofa_alloc_mem(ctrl_info, total_size, chunk_size)
[all...]
/linux-master/drivers/scsi/megaraid/
H A Dmegaraid_sas_fusion.c795 u32 chunk_size, array_size, offset; local
798 chunk_size = fusion->reply_alloc_sz * RDPQ_MAX_INDEX_IN_ONE_CHUNK;
816 chunk_size, 16, 0);
820 chunk_size,
821 roundup_pow_of_two(chunk_size),
863 chunk_size)) {
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c542 u32 xsk_chunk_size = rq->xsk_pool ? rq->xsk_pool->chunk_size : 0;
4422 max_mtu_frame = MLX5E_HW2SW_MTU(new_params, xsk.chunk_size - hr);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dsetup.c32 * and xsk->chunk_size is limited to 65535 bytes.
34 if ((size_t)xsk->chunk_size > PAGE_SIZE || xsk->chunk_size < MLX5E_MIN_XSK_CHUNK_SIZE) {
35 mlx5_core_err(mdev, "XSK chunk size %u out of bounds [%u, %lu]\n", xsk->chunk_size,
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dparams.c22 u8 req_page_shift = xsk ? order_base_2(xsk->chunk_size) : PAGE_SHIFT;
49 oversized = xsk->chunk_size < (1 << page_shift);
50 WARN_ON_ONCE(xsk->chunk_size > (1 << page_shift));
66 if (xsk->chunk_size % 3 == 0 && is_power_of_2(xsk->chunk_size / 3))
314 if (xsk && mlx5e_rx_get_linear_sz_xsk(params, xsk) > xsk->chunk_size)
569 1 << params->log_rq_mtu_frames, xsk->chunk_size);
1134 .chunk_size = 1 << frame_shift,
1148 xsk.chunk_size -= 1;
1153 xsk.chunk_size
[all...]
H A Dparams.h11 u16 chunk_size; member in struct:mlx5e_xsk_param
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_adminq_cmd.h1010 u8 chunk_size; member in struct:ice_aqc_layer_props
/linux-master/drivers/md/
H A Dmd.c1386 mddev->chunk_sectors = sb->chunk_size >> 9;
1562 sb->chunk_size = mddev->chunk_sectors << 9;
4339 __ATTR(chunk_size, S_IRUGO|S_IWUSR, chunk_size_show, chunk_size_store);
5285 /* Must be a multiple of chunk_size */
6787 info.chunk_size = mddev->chunk_sectors << 9;
7337 mddev->chunk_sectors = info->chunk_size >> 9;
7462 * The version, ctime,level,size,raid_disks,not_persistent, layout,chunk_size
7484 mddev->chunk_sectors != info->chunk_size >> 9 ||
H A Ddm-snap.c538 if (l->store->chunk_size < s->store->chunk_size)
858 unsigned int chunk_size = rounddown_pow_of_two(UINT_MAX); local
862 chunk_size = min_not_zero(chunk_size,
863 snap->store->chunk_size);
865 return (uint32_t) chunk_size;
1017 sector_t sector, unsigned int chunk_size);
1082 io_size = linear_chunks * s->store->chunk_size;
1378 s->store->chunk_size
[all...]
/linux-master/drivers/s390/cio/
H A Dcss.c1092 size_t chunk_size = chunk->end_addr - chunk->start_addr + 1; local
1094 dma_free_coherent((struct device *) data, chunk_size,
1121 size_t chunk_size; local
1128 chunk_size = round_up(size, PAGE_SIZE);
1129 addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP);
1132 gen_pool_add_virt(gp_dma, (unsigned long)addr, dma_addr, chunk_size, -1);
/linux-master/drivers/nvme/host/
H A Dpci.c1940 u32 chunk_size)
1949 tmp = (preferred + chunk_size - 1);
1950 do_div(tmp, chunk_size);
1968 len = min_t(u64, chunk_size, preferred - size);
2011 u64 chunk_size; local
2014 for (chunk_size = min_chunk; chunk_size >= hmminds; chunk_size /= 2) {
2015 if (!__nvme_alloc_host_mem(dev, preferred, chunk_size)) {
1939 __nvme_alloc_host_mem(struct nvme_dev *dev, u64 preferred, u32 chunk_size) argument
/linux-master/lib/
H A Dscatterlist.c514 unsigned int j, chunk_size; local
537 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset;
539 min_t(unsigned long, size, chunk_size), offset);
541 size -= chunk_size;
/linux-master/fs/nilfs2/
H A Ddir.c113 unsigned int chunk_size = nilfs_chunk_size(dir); local
121 if (limit & (chunk_size - 1))
136 if (((offs + rec_len - 1) ^ offs) & ~(chunk_size-1))
427 unsigned int chunk_size = nilfs_chunk_size(dir); local
456 rec_len = chunk_size;
457 de->rec_len = nilfs_rec_len_to_disk(chunk_size);
562 unsigned int chunk_size = nilfs_chunk_size(inode); local
570 err = nilfs_prepare_chunk(folio, 0, chunk_size);
576 memset(kaddr, 0, chunk_size);
586 de->rec_len = nilfs_rec_len_to_disk(chunk_size
[all...]
/linux-master/net/xdp/
H A Dxsk.c639 ts = pool->unaligned ? len : pool->chunk_size;
1333 __u32 chunk_size; member in struct:xdp_umem_reg_v1
1340 __u32 chunk_size; member in struct:xdp_umem_reg_v2
/linux-master/tools/testing/selftests/net/
H A Dtls.c434 uint16_t chunk_size,
447 EXPECT_GE(chunk_size, 1);
448 test_payload_size = chunk_size + extra_payload_size;
456 ret = sendfile(self->fd, fd, &offset, chunk_size);
432 chunked_sendfile(struct __test_metadata *_metadata, struct _test_data_tls *self, uint16_t chunk_size, uint16_t extra_payload_size) argument
/linux-master/tools/testing/selftests/mm/
H A Duffd-unit-tests.c1113 uffd_move_test_common(uffd_test_args_t *targs, unsigned long chunk_size, argument
1138 step_size = chunk_size / page_size;
1141 if (chunk_size > page_size) {
1142 char *aligned_src = ALIGN_UP(area_src, chunk_size);
1143 char *aligned_dst = ALIGN_UP(area_dst, chunk_size);
/linux-master/include/net/
H A Dxdp_sock.h27 u32 chunk_size; member in struct:xdp_umem
/linux-master/drivers/infiniband/hw/irdma/
H A Dverbs.c2820 stag_info->chunk_size = 1;
2823 stag_info->chunk_size = 3;
3560 stag_info.chunk_size = 1;
/linux-master/drivers/infiniband/hw/hns/
H A Dhns_roce_hem.c208 u32 chunk_size; local
224 chunk_size = table->type < HEM_TYPE_MTT ? mhop->buf_chunk_size :
226 table_idx = *obj / (chunk_size / table->obj_size);
/linux-master/arch/x86/xen/
H A Dsetup.c688 phys_addr_t mem_end, addr, size, chunk_size; local
773 chunk_size = size;
781 chunk_size = min(size, mem_end - addr);
783 chunk_size = min(size, PFN_PHYS(extra_pages));
785 n_pfns = PFN_DOWN(addr + chunk_size) - pfn_s;
794 xen_align_and_add_e820_region(addr, chunk_size, type);
796 addr += chunk_size;
797 size -= chunk_size;

Completed in 558 milliseconds

1234567