/linux-master/kernel/sched/ |
H A D | fair.c | 1287 __schedstat_set(tsk->stats.block_start,
|
H A D | core.c | 10283 schedstat_set(p->stats.block_start, 0);
|
/linux-master/fs/btrfs/tests/ |
H A D | extent-map-tests.c | 31 "em leak: em (start %llu len %llu block_start %llu block_len %llu) refs %d", 32 em->start, em->len, em->block_start, 79 em->block_start = 0; 100 em->block_start = SZ_32K; /* avoid merging */ 121 em->block_start = start; 137 em->block_start != 0 || em->block_len != SZ_16K) { 139 "case1 [%llu %llu]: ret %d return a wrong em (start %llu len %llu block_start %llu block_len %llu", 141 em->block_start, em->block_len); 175 em->block_start = EXTENT_MAP_INLINE; 196 em->block_start 466 add_compressed_extent(struct btrfs_fs_info *fs_info, struct extent_map_tree *em_tree, u64 start, u64 len, u64 block_start) argument [all...] |
H A D | inode-tests.c | 267 if (em->block_start != EXTENT_MAP_HOLE) { 268 test_err("expected a hole, got %llu", em->block_start); 286 if (em->block_start != EXTENT_MAP_INLINE) { 287 test_err("expected an inline, got %llu", em->block_start); 324 if (em->block_start != EXTENT_MAP_HOLE) { 325 test_err("expected a hole, got %llu", em->block_start); 347 if (em->block_start >= EXTENT_MAP_LAST_BYTE) { 348 test_err("expected a real extent, got %llu", em->block_start); 375 if (em->block_start >= EXTENT_MAP_LAST_BYTE) { 376 test_err("expected a real extent, got %llu", em->block_start); [all...] |
/linux-master/fs/btrfs/ |
H A D | inode.c | 141 u64 len, u64 orig_start, u64 block_start, 1204 ins.objectid, /* block_start */ 1282 if (em->block_start >= EXTENT_MAP_LAST_BYTE) { 1285 if (em && em->block_start < EXTENT_MAP_LAST_BYTE) 1286 alloc_hint = em->block_start; 1290 alloc_hint = em->block_start; 1440 ins.objectid, /* block_start */ 2169 nocow_args.disk_bytenr, /* block_start */ 2684 if (em->block_start != EXTENT_MAP_HOLE) 3242 * writepage where we do ASSERT(em->block_start ! 4744 u64 block_start; local 7006 btrfs_create_dio_extent(struct btrfs_inode *inode, struct btrfs_dio_data *dio_data, const u64 start, const u64 len, const u64 orig_start, const u64 block_start, const u64 block_len, const u64 orig_block_len, const u64 ram_bytes, const int type) argument 7310 create_io_em(struct btrfs_inode *inode, u64 start, u64 len, u64 orig_start, u64 block_start, u64 block_len, u64 orig_block_len, u64 ram_bytes, int compress_type, int type) argument 7363 u64 block_start, orig_start, orig_block_len, ram_bytes; local 10627 u64 block_start; member in struct:btrfs_swap_info [all...] |
H A D | extent_map.c | 191 if (em->block_start + em->block_len < em->block_start) 193 return em->block_start + em->block_len; 228 if (next->block_start < EXTENT_MAP_LAST_BYTE - 1) 229 return next->block_start == extent_map_block_end(prev); 232 return next->block_start == prev->block_start; 264 em->block_start = merge->block_start; 564 if (em->block_start < EXTENT_MAP_LAST_BYT [all...] |
H A D | extent_io.c | 1033 u64 block_start; 1086 disk_bytenr = em->block_start; 1088 disk_bytenr = em->block_start + extent_offset; 1089 block_start = em->block_start; 1091 block_start = EXTENT_MAP_HOLE; 1139 if (block_start == EXTENT_MAP_HOLE) { 1149 if (block_start == EXTENT_MAP_INLINE) { 1351 u64 block_start; 1408 block_start 1002 u64 block_start; local 1320 u64 block_start; local [all...] |
H A D | zoned.c | 1772 em->block_start = logical;
|
H A D | tree-log.c | 4587 em->block_start == EXTENT_MAP_HOLE) 4658 csum_root = btrfs_csum_root(trans->fs_info, em->block_start); 4659 ret = btrfs_lookup_csums_list(csum_root, em->block_start + csum_offset, 4660 em->block_start + csum_offset + 4704 btrfs_set_stack_file_extent_disk_bytenr(&fi, em->block_start); 4706 } else if (em->block_start < EXTENT_MAP_LAST_BYTE) { 4707 btrfs_set_stack_file_extent_disk_bytenr(&fi, em->block_start -
|
H A D | relocation.c | 2903 u64 start, u64 end, u64 block_start) 2916 em->block_start = block_start; 2940 setup_relocation_extent_mapping(struct inode *inode, u64 start, u64 end, u64 block_start) argument
|
H A D | file.c | 101 u64 block_start = round_down(pos, fs_info->sectorsize); local 102 u64 block_len = round_up(pos + copied, fs_info->sectorsize) - block_start; 113 block_start, block_len); 2339 hole_em->block_start = EXTENT_MAP_HOLE; 2372 if (em->block_start == EXTENT_MAP_HOLE) { 3029 if (em->block_start == EXTENT_MAP_HOLE) 3093 alloc_hint = em->block_start + em->len; 3111 if (len < sectorsize && em->block_start != EXTENT_MAP_HOLE) { 3324 if (em->block_start == EXTENT_MAP_HOLE ||
|
H A D | file-item.c | 1300 em->block_start = EXTENT_MAP_HOLE; 1305 em->block_start = bytenr; 1309 em->block_start = bytenr; 1318 em->block_start = EXTENT_MAP_INLINE;
|
H A D | extent_map.h | 95 u64 block_start; 54 u64 block_start; member in struct:extent_map
|
H A D | defrag.c | 711 em->block_start = EXTENT_MAP_HOLE; 828 if (!next || next->block_start >= EXTENT_MAP_LAST_BYTE) 995 if (em->block_start == EXTENT_MAP_INLINE && 1000 if (em->block_start == EXTENT_MAP_HOLE || 1065 if (em->block_start == EXTENT_MAP_INLINE)
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.c | 512 u64 block_start = round_up(offset, block_bytes(c)); local 525 if (block_start < block_end) { 529 block_start >> 9, block_end >> 9, 709 u64 block_start = round_down(offset, block_bytes(c)); local 729 block_start = round_up(offset, block_bytes(c)); 733 ret = __bchfs_fallocate(inode, mode, block_start >> 9, block_end >> 9);
|
/linux-master/fs/cramfs/ |
H A D | inode.c | 827 u32 block_ptr, block_start, block_len; local 843 block_start = block_ptr << CRAMFS_BLK_DIRECT_PTR_SHIFT; 852 cramfs_read(sb, block_start, 2); 853 block_start += 2; 863 block_start = OFFSET(inode) + maxblock * 4; 865 block_start = *(u32 *) 868 if (unlikely(block_start & CRAMFS_BLK_FLAG_DIRECT_PTR)) { 870 u32 prev_start = block_start; 871 block_start = prev_start & ~CRAMFS_BLK_FLAGS; 872 block_start << [all...] |
/linux-master/tools/net/ynl/ |
H A D | ynl-gen-c.py | 197 ri.cw.block_start(line=f"{kw} (type == {self.enum_name})") 1264 def block_start(self, line=''): 1353 self.block_start() 1515 cw.block_start(line=f'struct ynl_policy_attr {struct.render_name}_policy[{type_max} + 1] =') 1523 cw.block_start(line=f'struct ynl_policy_nest {struct.render_name}_nest =') 1535 cw.block_start() 1551 cw.block_start(line=f"static const char * const {map_name}[] =") 1577 cw.block_start(line=f"static const char * const {map_name}[] =") 1597 ri.cw.block_start() 1645 ri.cw.block_start() 1261 def block_start(self, line=''): member in class:CodeWriter [all...] |
/linux-master/include/linux/ |
H A D | sched.h | 507 u64 block_start; member in struct:sched_statistics
|
/linux-master/fs/ext4/ |
H A D | inode.c | 955 unsigned block_start, block_end; local 960 for (bh = head, block_start = 0; 961 ret == 0 && (bh != head || !block_start); 962 block_start = block_end, bh = next) { 964 block_end = block_start + blocksize; 965 if (block_end <= from || block_start >= to) { 1022 unsigned block_start, block_end; local 1042 for (bh = head, block_start = 0; bh != head || !block_start; 1043 block++, block_start 1348 unsigned int block_start = 0, block_end; local [all...] |
/linux-master/fs/ocfs2/ |
H A D | file.c | 762 unsigned zero_from, zero_to, block_start, block_end; local 797 for (block_start = zero_from; block_start < zero_to; 798 block_start = block_end) { 799 block_end = block_start + i_blocksize(inode); 802 * block_start is block-aligned. Bump it by one to force 806 ret = __block_write_begin(page, block_start + 1, 0, 815 block_commit_write(page, block_start + 1, block_start + 1);
|
/linux-master/fs/nilfs2/ |
H A D | page.c | 438 unsigned int block_start, block_end; local 442 for (bh = head = page_buffers(page), block_start = 0; 443 bh != head || !block_start; 444 block_start = block_end, bh = bh->b_this_page) { 445 block_end = block_start + bh->b_size; 446 if (block_end > from && block_start < to && !buffer_dirty(bh))
|
/linux-master/drivers/md/ |
H A D | dm-writecache.c | 153 void *block_start; member in struct:dm_writecache 400 return (char *)wc->block_start + (e->index << wc->block_size_bits); 2643 wc->block_start = (char *)sb(wc) + offset;
|
/linux-master/arch/arm/mm/ |
H A D | mmu.c | 1198 phys_addr_t block_start, block_end, memblock_limit = 0; local 1216 for_each_mem_range(i, &block_start, &block_end) { 1217 if (!IS_ALIGNED(block_start, PMD_SIZE)) { 1220 len = round_up(block_start, PMD_SIZE) - block_start; 1221 memblock_mark_nomap(block_start, len); 1226 for_each_mem_range(i, &block_start, &block_end) { 1227 if (block_start < vmalloc_limit) { 1253 if (!IS_ALIGNED(block_start, PMD_SIZE)) 1254 memblock_limit = block_start; [all...] |
/linux-master/fs/iomap/ |
H A D | buffered-io.c | 659 static int iomap_read_folio_sync(loff_t block_start, struct folio *folio, argument 666 bio.bi_iter.bi_sector = iomap_sector(iomap, block_start); 677 loff_t block_start = round_down(pos, block_size); local 703 iomap_adjust_read_range(iter->inode, folio, &block_start, 704 block_end - block_start, &poff, &plen); 713 if (iomap_block_needs_zeroing(iter, block_start)) { 723 status = iomap_read_folio_sync(block_start, folio, 729 } while ((block_start += plen) < block_end);
|
/linux-master/fs/ |
H A D | buffer.c | 1966 size_t block_start, block_end; local 1975 block_start = 0; 1977 block_end = block_start + bh->b_size; 1980 if (block_end > from && block_start < to) { 1984 start = max(from, block_start); 1996 block_start = block_end; 2076 size_t block_start, block_end; local 2090 for (bh = head, block_start = 0; bh != head || !block_start; 2091 block++, block_start 2161 size_t block_start, block_end; local 2307 unsigned block_start, block_end, blocksize; local [all...] |