/linux-master/mm/ |
H A D | usercopy.c | 167 struct folio *folio; local 192 folio = virt_to_folio(ptr); 194 if (folio_test_slab(folio)) { 196 __check_heap_object(ptr, n, folio_slab(folio), to_user); 197 } else if (folio_test_large(folio)) { 198 offset = ptr - folio_address(folio); 199 if (n > folio_size(folio) - offset)
|
H A D | mmzone.c | 97 int folio_xchg_last_cpupid(struct folio *folio, int cpupid) argument 102 old_flags = READ_ONCE(folio->flags); 109 } while (unlikely(!try_cmpxchg(&folio->flags, &old_flags, flags)));
|
H A D | userfaultfd.c | 185 struct folio *folio; local 215 folio = page_folio(page); 219 folio_add_lru(folio); 220 folio_add_file_rmap_pte(folio, page, dst_vma); 222 folio_add_new_anon_rmap(folio, dst_vma, dst_addr); 223 folio_add_lru_vma(folio, dst_vma); 230 inc_mm_counter(dst_mm, mm_counter(folio)); 248 struct folio **foliop) 252 struct folio *foli local 358 struct folio *folio; local 472 struct folio *folio; local 680 struct folio *folio; local 1188 struct folio *folio; local 1665 struct folio *folio = pfn_folio(pmd_pfn(*src_pmd)); local [all...] |
H A D | khugepaged.c | 495 static void release_pte_folio(struct folio *folio) argument 497 node_stat_mod_folio(folio, 498 NR_ISOLATED_ANON + folio_is_file_lru(folio), 499 -folio_nr_pages(folio)); 500 folio_unlock(folio); 501 folio_putback_lru(folio); 507 struct folio *folio, *tmp; local 518 folio 530 is_refcount_suitable(struct folio *folio) argument 548 struct folio *folio = NULL; local 650 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 894 hpage_collapse_alloc_folio(struct folio **folio, gfp_t gfp, int node, nodemask_t *nmask) argument 1068 struct folio *folio; local 1095 struct folio *folio; local 1261 struct folio *folio = NULL; local 1487 struct folio *folio; local 1803 struct folio *folio; local [all...] |
H A D | memcontrol.c | 365 * mem_cgroup_css_from_folio - css of the memcg associated with a folio 366 * @folio: folio of interest 369 * with @folio is returned. The returned css remains associated with @folio 375 struct cgroup_subsys_state *mem_cgroup_css_from_folio(struct folio *folio) argument 377 struct mem_cgroup *memcg = folio_memcg(folio); 900 void __lruvec_stat_mod_folio(struct folio *folio, enu argument 1343 lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) argument 1355 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio); local 1371 folio_lruvec_lock(struct folio *folio) argument 1394 folio_lruvec_lock_irq(struct folio *folio) argument 1418 folio_lruvec_lock_irqsave(struct folio *folio, unsigned long *flags) argument 2185 folio_memcg_lock(struct folio *folio) argument 2251 folio_memcg_unlock(struct folio *folio) argument 2949 commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 2951 VM_BUG_ON_FOLIO(folio_memcg(folio), folio); local 2969 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 3043 mem_cgroup_from_obj_folio(struct folio *folio, void *p) argument 3093 struct folio *folio; local 3232 get_obj_cgroup_from_folio(struct folio *folio) argument 3346 struct folio *folio = page_folio(page); local 3612 struct folio *folio = page_folio(head); local 4868 mem_cgroup_track_foreign_dirty_slowpath(struct folio *folio, struct bdi_writeback *wb) argument 5879 struct folio *folio; member in union:mc_target 5954 struct folio *folio; local 5982 mem_cgroup_move_account(struct folio *folio, bool compound, struct mem_cgroup *from, struct mem_cgroup *to) argument 5994 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 6115 struct folio *folio; local 6200 struct folio *folio; local 6441 struct folio *folio; local 7279 charge_memcg(struct folio *folio, struct mem_cgroup *memcg, gfp_t gfp) argument 7293 __mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp) argument 7352 mem_cgroup_swapin_charge_folio(struct folio *folio, struct mm_struct *mm, gfp_t gfp, swp_entry_t entry) argument 7444 uncharge_folio(struct folio *folio, struct uncharge_gather *ug) argument 7450 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 7503 __mem_cgroup_uncharge(struct folio *folio) argument 7797 mem_cgroup_swapout(struct folio *folio, swp_entry_t entry) argument 7803 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 7804 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local 7867 __mem_cgroup_try_charge_swap(struct folio *folio, swp_entry_t entry) argument 7947 mem_cgroup_swap_full(struct folio *folio) argument [all...] |
/linux-master/mm/damon/ |
H A D | ops-common.h | 10 struct folio *damon_get_folio(unsigned long pfn);
|
/linux-master/fs/nilfs2/ |
H A D | mdt.c | 360 struct folio *folio; local 365 folio = filemap_lock_folio(inode->i_mapping, index); 366 if (IS_ERR(folio)) 369 folio_wait_writeback(folio); 371 bh = folio_buffers(folio); 378 still_dirty = folio_test_dirty(folio); 379 folio_unlock(folio); 380 folio_put(folio); 402 struct folio *foli local 562 struct folio *folio; local 596 struct folio *folio; local [all...] |
/linux-master/arch/microblaze/include/asm/ |
H A D | cacheflush.h | 77 static inline void flush_dcache_folio(struct folio *folio) argument 79 unsigned long addr = folio_pfn(folio) << PAGE_SHIFT; 81 flush_dcache_range(addr, addr + folio_size(folio));
|
/linux-master/fs/ocfs2/ |
H A D | symlink.c | 55 static int ocfs2_fast_symlink_read_folio(struct file *f, struct folio *folio) argument 57 struct page *page = &folio->page;
|
/linux-master/fs/efs/ |
H A D | symlink.c | 15 static int efs_symlink_read_folio(struct file *file, struct folio *folio) argument 17 struct page *page = &folio->page;
|
/linux-master/arch/sh/mm/ |
H A D | kmap.c | 30 struct folio *folio = page_folio(page); local 34 BUG_ON(!test_bit(PG_dcache_clean, &folio->flags));
|
/linux-master/fs/btrfs/ |
H A D | verity.c | 717 struct folio *folio; local 728 folio = __filemap_get_folio(inode->i_mapping, index, FGP_ACCESSED, 0); 729 if (!IS_ERR(folio)) { 730 if (folio_test_uptodate(folio)) 733 folio_lock(folio); 735 if (!folio_test_uptodate(folio)) { 736 folio_unlock(folio); 737 folio_put(folio); 740 folio_unlock(folio); [all...] |
/linux-master/include/linux/ |
H A D | iomap.h | 133 * and put_folio will be called for each folio written to. This only applies 139 * @folio. 142 struct folio *(*get_folio)(struct iomap_iter *iter, loff_t pos, 145 struct folio *folio); 159 * This is called with the folio over the specified file position held 265 int iomap_read_folio(struct folio *folio, const struct iomap_ops *ops); 267 bool iomap_is_partially_uptodate(struct folio *, size_t from, size_t count); 268 struct folio *iomap_get_foli [all...] |
H A D | highmem-internal.h | 76 static inline void *kmap_local_folio(struct folio *folio, size_t offset) argument 78 struct page *page = folio_page(folio, offset / PAGE_SIZE); 188 static inline void *kmap_local_folio(struct folio *folio, size_t offset) argument 190 return page_address(&folio->page) + offset;
|
H A D | iov_iter.h | 150 struct folio *folio; local 157 xas_for_each(&xas, folio, ULONG_MAX) { 160 if (xas_retry(&xas, folio)) 162 if (WARN_ON(xa_is_value(folio))) 164 if (WARN_ON(folio_test_hugetlb(folio))) 167 offset = offset_in_folio(folio, start + progress); 168 flen = min(folio_size(folio) - offset, len); 171 void *base = kmap_local_folio(folio, offset);
|
H A D | dax.h | 163 dax_entry_t dax_lock_folio(struct folio *folio); 164 void dax_unlock_folio(struct folio *folio, dax_entry_t cookie); 186 static inline dax_entry_t dax_lock_folio(struct folio *folio) argument 188 if (IS_DAX(folio->mapping->host)) 193 static inline void dax_unlock_folio(struct folio *folio, dax_entry_t cookie) argument
|
/linux-master/fs/ext4/ |
H A D | inline.c | 480 static int ext4_read_inline_folio(struct inode *inode, struct folio *folio) argument 487 BUG_ON(!folio_test_locked(folio)); 489 BUG_ON(folio->index); 503 kaddr = kmap_local_folio(folio, 0); 505 kaddr = folio_zero_tail(folio, len, kaddr + len); 507 folio_mark_uptodate(folio); 514 int ext4_readpage_inline(struct inode *inode, struct folio *folio) argument 528 if (!folio 547 struct folio *folio = NULL; local 667 struct folio *folio; local 742 ext4_write_inline_data_end(struct inode *inode, loff_t pos, unsigned len, unsigned copied, struct folio *folio) argument 838 struct folio *folio; local 899 struct folio *folio; local [all...] |
H A D | inode.c | 978 * Helper for handling dirtying of journalled data. We also mark the folio as 1016 static int ext4_block_write_begin(struct folio *folio, loff_t pos, unsigned len, argument 1021 struct inode *inode = folio->mapping->host; 1031 BUG_ON(!folio_test_locked(folio)); 1036 head = folio_buffers(folio); 1038 head = create_empty_buffers(folio, blocksize, 0); 1040 block = (sector_t)folio->index << (PAGE_SHIFT - bbits); 1046 if (folio_test_uptodate(folio)) { 1059 if (folio_test_uptodate(folio)) { 1125 struct folio *folio; local 1274 struct folio *folio = page_folio(page); local 1343 ext4_journalled_zero_new_buffers(handle_t *handle, struct inode *inode, struct folio *folio, unsigned from, unsigned to) argument 1378 struct folio *folio = page_folio(page); local 1580 struct folio *folio = fbatch.folios[i]; local 1837 mpage_folio_done(struct mpage_da_data *mpd, struct folio *folio) argument 1843 mpage_submit_folio(struct mpage_da_data *mpd, struct folio *folio) argument 2010 mpage_process_folio(struct mpage_da_data *mpd, struct folio *folio, ext4_lblk_t *m_lblk, ext4_fsblk_t *m_pblk, bool *map_bh) argument 2103 struct folio *folio = fbatch.folios[i]; local 2303 ext4_journal_folio_buffers(handle_t *handle, struct folio *folio, size_t len) argument 2324 mpage_journal_page_buffers(handle_t *handle, struct mpage_da_data *mpd, struct folio *folio) argument 2398 struct folio *folio = fbatch.folios[i]; local 2858 struct folio *folio; local 2923 ext4_da_should_update_i_disksize(struct folio *folio, unsigned long offset) argument 2942 ext4_da_do_write_end(struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct folio *folio) argument 3011 struct folio *folio = page_folio(page); local 3119 ext4_read_folio(struct file *file, struct folio *folio) argument 3146 ext4_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 3157 __ext4_journalled_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 3174 ext4_journalled_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 3181 ext4_release_folio(struct folio *folio, gfp_t wait) argument 3500 ext4_journalled_dirty_folio(struct address_space *mapping, struct folio *folio) argument 3509 ext4_dirty_folio(struct address_space *mapping, struct folio *folio) argument 3617 struct folio *folio; local 5226 struct folio *folio = filemap_lock_folio(inode->i_mapping, local 6044 struct folio *folio = page_folio(vmf->page); local [all...] |
/linux-master/fs/f2fs/ |
H A D | verity.c | 261 struct folio *folio; local 265 folio = __filemap_get_folio(inode->i_mapping, index, FGP_ACCESSED, 0); 266 if (IS_ERR(folio) || !folio_test_uptodate(folio)) { 269 if (!IS_ERR(folio)) 270 folio_put(folio); 273 folio = read_mapping_folio(inode->i_mapping, index, NULL); 274 if (IS_ERR(folio)) 275 return ERR_CAST(folio); [all...] |
/linux-master/fs/gfs2/ |
H A D | meta_io.c | 118 struct folio *folio; local 132 folio = __filemap_get_folio(mapping, index, 135 bh = folio_buffers(folio); 137 bh = create_empty_buffers(folio, 140 folio = __filemap_get_folio(mapping, index, 142 if (IS_ERR(folio)) 144 bh = folio_buffers(folio); 155 folio_unlock(folio); 156 folio_put(folio); 406 struct folio *folio; local [all...] |
/linux-master/drivers/scsi/ |
H A D | scsicam.c | 37 struct folio *folio; local 39 folio = read_mapping_folio(mapping, 0, NULL); 40 if (IS_ERR(folio)) 43 res = kmemdup(folio_address(folio) + 0x1be, 66, GFP_KERNEL); 44 folio_put(folio);
|
/linux-master/fs/bcachefs/ |
H A D | fs-io.h | 15 struct folio *fv_folio; 23 struct folio *folio = page_folio(bv.bv_page); local 24 size_t offset = (folio_page_idx(folio, bv.bv_page) << PAGE_SHIFT) + 26 size_t len = min_t(size_t, folio_size(folio) - offset, bv.bv_len); 29 .fv_folio = folio,
|
/linux-master/fs/xfs/ |
H A D | xfs_buf_mem.c | 136 struct folio *folio = NULL; local 152 error = shmem_get_folio(inode, pos >> PAGE_SHIFT, &folio, SGP_CACHE); 157 folio_unlock(folio); 158 folio_put(folio); 162 page = folio_file_page(folio, pos >> PAGE_SHIFT);
|
/linux-master/fs/squashfs/ |
H A D | symlink.c | 33 static int squashfs_symlink_read_folio(struct file *file, struct folio *folio) argument 35 struct page *page = &folio->page;
|
/linux-master/arch/riscv/mm/ |
H A D | cacheflush.c | 87 struct folio *folio = page_folio(pte_page(pte)); local 89 if (!test_bit(PG_dcache_clean, &folio->flags)) { 91 set_bit(PG_dcache_clean, &folio->flags);
|