/linux-master/tools/testing/selftests/mm/ |
H A D | vm_util.c | 20 uint64_t pagemap_get_entry(int fd, char *start) argument 22 const unsigned long pfn = (unsigned long)start / getpagesize(); 32 static uint64_t __pagemap_scan_get_categories(int fd, char *start, struct page_region *r) argument 36 arg.start = (uintptr_t)start; 37 arg.end = (uintptr_t)(start + psize()); 53 static uint64_t pagemap_scan_get_categories(int fd, char *start) argument 58 ret = __pagemap_scan_get_categories(fd, start, &r); 66 /* `start` is any valid address. */ 67 static bool pagemap_scan_supported(int fd, char *start) argument 85 page_entry_is(int fd, char *start, char *desc, uint64_t pagemap_flags, uint64_t pagescan_flags) argument 102 pagemap_is_softdirty(int fd, char *start) argument 108 pagemap_is_swapped(int fd, char *start) argument 113 pagemap_is_populated(int fd, char *start) argument 120 pagemap_get_pfn(int fd, char *start) argument [all...] |
/linux-master/fs/btrfs/ |
H A D | extent_io.h | 83 #define BITMAP_FIRST_BYTE_MASK(start) \ 84 ((BYTE_MASK << ((start) & (BITS_PER_BYTE - 1))) & BYTE_MASK) 94 u64 start; member in struct:extent_buffer 136 u64 start) 139 return start & (eb->folio_size - 1); 146 * @start: offset inside the extent buffer 156 * The eb->start is aligned to folio size, thus adding it 159 * The eb->start is aligned to folio (page) size, thus 165 * We need to add eb->start to properly access the offset inside 168 return offset_in_folio(eb->folios[0], offset + eb->start); 135 offset_in_eb_folio(const struct extent_buffer *eb, u64 start) argument [all...] |
H A D | misc.h | 62 * structure start with these members. 93 * Return the rb_node that start at or after @bytenr. If there is no entry at 147 unsigned long start, 152 found_zero = find_next_zero_bit(addr, start + nbits, start); 153 return (found_zero == start + nbits); 157 unsigned long start, 162 found_set = find_next_bit(addr, start + nbits, start); 163 return (found_set == start 146 bitmap_test_range_all_set(const unsigned long *addr, unsigned long start, unsigned long nbits) argument 156 bitmap_test_range_all_zero(const unsigned long *addr, unsigned long start, unsigned long nbits) argument [all...] |
H A D | subpage.c | 225 struct folio *folio, u64 start, u32 len) 232 ASSERT(IS_ALIGNED(start, fs_info->sectorsize) && 239 ASSERT(folio_pos(folio) <= start && 240 start + len <= folio_pos(folio) + PAGE_SIZE); 243 #define subpage_calc_start_bit(fs_info, folio, name, start, len) \ 247 btrfs_subpage_assert(fs_info, folio, start, len); \ 248 start_bit = offset_in_page(start) >> fs_info->sectorsize_bits; \ 254 struct folio *folio, u64 start, u32 len) 257 const int start_bit = subpage_calc_start_bit(fs_info, folio, locked, start, len); 262 btrfs_subpage_assert(fs_info, folio, start, le 224 btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 253 btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 275 btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 309 btrfs_subpage_clamp_range(struct folio *folio, u64 *start, u32 *len) argument 327 btrfs_subpage_start_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 347 btrfs_subpage_end_and_test_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 390 btrfs_folio_start_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 407 btrfs_folio_end_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 429 btrfs_subpage_set_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 444 btrfs_subpage_clear_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 458 btrfs_subpage_set_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 482 btrfs_subpage_clear_and_test_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 499 btrfs_subpage_clear_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 509 btrfs_subpage_set_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 524 btrfs_subpage_clear_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 541 btrfs_subpage_set_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 555 btrfs_subpage_clear_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 570 btrfs_subpage_set_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 585 btrfs_subpage_clear_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 736 btrfs_folio_unlock_writer(struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 772 btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument [all...] |
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_book3s_64.c | 24 static void __init kasan_init_phys_region(void *start, void *end) argument 29 if (start >= end) 32 k_start = ALIGN_DOWN((unsigned long)kasan_mem_to_shadow(start), PAGE_SIZE); 50 * This takes us up to where we start at c00e... 55 phys_addr_t start, end; local 64 for_each_mem_range(i, &start, &end) 65 kasan_init_phys_region(phys_to_virt(start), phys_to_virt(end));
|
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_pkvm.h | 30 unsigned long start, end; local 32 start = (reg->base >> PAGE_SHIFT) * vmemmap_entry_size; 33 end = start + nr_pages * vmemmap_entry_size; 34 start = ALIGN_DOWN(start, PAGE_SIZE); 37 return end - start;
|
/linux-master/arch/mips/jazz/ |
H A D | reset.c | 40 unsigned long start = jiffies; local 41 unsigned long timeout = start + HZ/2;
|
/linux-master/arch/mips/cobalt/ |
H A D | pci.c | 20 .start = GT_DEF_PCI0_MEM0_BASE, 27 .start = 0x1000,
|
/linux-master/fs/qnx4/ |
H A D | bitmap.c | 23 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; local 33 if ((bh = sb_bread(sb, start + offset)) == NULL) {
|
/linux-master/arch/nios2/lib/ |
H A D | delay.c | 13 cycles_t start = get_cycles(); local 15 while ((get_cycles() - start) < cycles)
|
/linux-master/include/linux/ |
H A D | irqbypass.h | 24 * the @add_* callbacks, and finally the optional @start callback will allow 39 * @start: Perform any startup operations necessary after add/del (optional) 54 void (*start)(struct irq_bypass_producer *); member in struct:irq_bypass_producer 64 * @start: Perform any startup operations necessary after add/del (optional) 79 void (*start)(struct irq_bypass_consumer *); member in struct:irq_bypass_consumer
|
/linux-master/arch/m68k/include/asm/ |
H A D | cacheflush_no.h | 12 #define flush_dcache_range(start, len) __flush_dcache_all() 13 #define flush_icache_range(start, len) __flush_icache_all()
|
/linux-master/drivers/staging/media/atomisp/pci/mmu/ |
H A D | isp_mmu.c | 49 * physical address 0 - 0x7fffffff and the rest will start from 0x100000000. 201 unsigned int start, unsigned int end, phys_addr_t phys) 209 start = start & ISP_PAGE_MASK; 213 ptr = start; 224 free_mmu_map(mmu, start, ptr); 245 unsigned int start, unsigned int end, 256 start = start & ISP_PAGE_MASK; 260 ptr = start; 199 mmu_l2_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument 244 mmu_l1_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument 320 unsigned int start, end; local 363 mmu_l2_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end) argument 401 mmu_l1_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end) argument 451 unsigned int start, end; local 478 unsigned int start, end; local 498 isp_mmu_flush_tlb_range_default(struct isp_mmu *mmu, unsigned int start, unsigned int size) argument [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | mmzone.h | 28 void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end); 31 setup_bootmem_node(int nid, unsigned long start, unsigned long end) argument
|
/linux-master/drivers/pci/ |
H A D | host-bridge.c | 65 region->start = res->start - offset; 73 return region1->start <= region2->start && region1->end >= region2->end; 89 bus_region.start = window->res->start - window->offset; 98 res->start = region->start + offset;
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_gtt.c | 113 node->start = offset; 130 static u64 random_offset(u64 start, u64 end, u64 len, u64 align) argument 134 GEM_BUG_ON(range_overflows(start, len, end)); 135 GEM_BUG_ON(round_up(start, align) > round_down(end - len, align)); 137 range = round_down(end - len, align) - round_up(start, align); 149 start += addr; 152 return round_up(start, align); 166 * @start: start of any range restriction inside GTT (0 for all), 174 * its @size must then fit entirely within the [@start, 190 i915_gem_gtt_insert(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_mm_node *node, u64 size, u64 alignment, unsigned long color, u64 start, u64 end, unsigned int flags) argument [all...] |
/linux-master/fs/squashfs/ |
H A D | xattr.h | 18 u64 start, u64 *xattr_table_start, unsigned int *xattr_ids) 22 id_table = squashfs_read_table(sb, start, sizeof(*id_table)); 17 squashfs_read_xattr_id_table(struct super_block *sb, u64 start, u64 *xattr_table_start, unsigned int *xattr_ids) argument
|
/linux-master/arch/arm/include/asm/ |
H A D | unwind.h | 28 const struct unwind_idx *start; member in struct:unwind_table 35 extern struct unwind_table *unwind_table_add(unsigned long start,
|
/linux-master/drivers/pinctrl/nxp/ |
H A D | pinctrl-s32.h | 29 * @start: start pin ID 33 unsigned int start; member in struct:s32_pin_range 55 #define S32_PIN_RANGE(_start, _end) { .start = _start, .end = _end }
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_migrate.h | 44 unsigned long start, unsigned long last, 48 unsigned long start, unsigned long last,
|
/linux-master/arch/mips/include/asm/ |
H A D | cacheflush.h | 22 * - flush_cache_range(vma, start, end) flushes a range of pages 23 * - flush_icache_range(start, end) flush a range of instructions 30 * - __flush_icache_user_range(start, end) flushes range of user instructions 51 unsigned long start, unsigned long end); 85 extern void (*flush_icache_range)(unsigned long start, unsigned long end); 86 extern void (*local_flush_icache_range)(unsigned long start, unsigned long end); 87 extern void (*__flush_icache_user_range)(unsigned long start, 89 extern void (*__local_flush_icache_user_range)(unsigned long start, 94 static inline void flush_cache_vmap(unsigned long start, unsigned long end) argument 100 #define flush_cache_vmap_early(start, en 104 flush_cache_vunmap(unsigned long start, unsigned long end) argument [all...] |
/linux-master/scripts/kconfig/ |
H A D | mnconf-common.c | 21 int handle_search_keys(int key, size_t start, size_t end, void *_data) argument 33 if (pos->offset < start)
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_range_fence.c | 13 #define XE_RANGE_TREE_START(_node) ((_node)->start) 48 * @start: start address of range fence 57 u64 start, u64 last, struct dma_fence *fence) 67 rfence->start = start; 128 * @start: start address of range fence 134 xe_range_fence_tree_first(struct xe_range_fence_tree *tree, u64 start, argument 137 return xe_range_fence_tree_iter_first(&tree->root, start, las 54 xe_range_fence_insert(struct xe_range_fence_tree *tree, struct xe_range_fence *rfence, const struct xe_range_fence_ops *ops, u64 start, u64 last, struct dma_fence *fence) argument 149 xe_range_fence_tree_next(struct xe_range_fence *rfence, u64 start, u64 last) argument [all...] |
/linux-master/arch/sh/kernel/ |
H A D | setup.c | 120 unsigned long start, end; local 129 start = INITRD_START + __MEMORY_START; 130 end = start + INITRD_SIZE; 132 if (unlikely(end <= start)) 134 if (unlikely(start & ~PAGE_MASK)) { 139 if (unlikely(start < __MEMORY_START)) { 140 pr_err("initrd start (%08lx) < __MEMORY_START(%x)\n", 141 start, __MEMORY_START); 162 initrd_start = (unsigned long)__va(start); 197 unsigned long start, en local [all...] |
/linux-master/drivers/video/ |
H A D | screen_info_generic.c | 9 resource_size_t start, resource_size_t size, 14 r->start = start; 15 r->end = start + size - 1; 21 resource_size_t start, resource_size_t size, 24 resource_init_named(r, start, size, name, IORESOURCE_IO); 28 resource_size_t start, resource_size_t size, 31 resource_init_named(r, start, size, name, IORESOURCE_MEM); 8 resource_init_named(struct resource *r, resource_size_t start, resource_size_t size, const char *name, unsigned int flags) argument 20 resource_init_io_named(struct resource *r, resource_size_t start, resource_size_t size, const char *name) argument 27 resource_init_mem_named(struct resource *r, resource_size_t start, resource_size_t size, const char *name) argument
|