/linux-master/mm/ |
H A D | zswap.c | 768 void zswap_folio_swapin(struct folio *folio) argument 772 if (folio) { 773 lruvec = folio_lruvec(folio); 902 static bool zswap_compress(struct folio *folio, struct zswap_entry *entry) argument 920 sg_set_page(&input, &folio->page, PAGE_SIZE, 0); 1017 * Attempts to free an entry by adding a folio to the swap cache, 1018 * decompressing the entry data into the folio, and issuing a 1019 * bio write to write the folio bac 1033 struct folio *folio; local 1376 zswap_is_folio_same_filled(struct folio *folio, unsigned long *value) argument 1412 zswap_store(struct folio *folio) argument 1550 zswap_load(struct folio *folio) argument [all...] |
H A D | hugetlb_cgroup.c | 194 struct folio *folio = page_folio(page); local 196 page_hcg = hugetlb_cgroup_from_folio(folio); 215 set_hugetlb_cgroup(folio, parent); 307 struct folio *folio, bool rsvd) 312 __set_hugetlb_cgroup(folio, h_cg, rsvd); 315 h_cg->nodeinfo[folio_nid(folio)]->usage[idx]; 321 WRITE_ONCE(h_cg->nodeinfo[folio_nid(folio)]->usage[idx], 328 struct folio *foli 305 __hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio, bool rsvd) argument 326 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 333 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 343 __hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio, bool rsvd) argument 375 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument 381 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument [all...] |
H A D | mincore.c | 55 struct folio *folio; local 63 folio = filemap_get_incore_folio(mapping, index); 64 if (!IS_ERR(folio)) { 65 present = folio_test_uptodate(folio); 66 folio_put(folio);
|
/linux-master/fs/nfs/ |
H A D | dir.c | 58 static void nfs_readdir_clear_array(struct folio *); 159 struct folio *folio; member in struct:nfs_readdir_descriptor 203 static void nfs_readdir_folio_init_array(struct folio *folio, u64 last_cookie, argument 208 array = kmap_local_folio(folio, 0); 221 static void nfs_readdir_clear_array(struct folio *folio) argument 226 array = kmap_local_folio(folio, 0); 233 static void nfs_readdir_folio_reinit_array(struct folio *foli argument 243 struct folio *folio = folio_alloc(gfp_flags, 0); local 249 nfs_readdir_folio_array_free(struct folio *folio) argument 311 nfs_readdir_folio_array_append(struct folio *folio, const struct nfs_entry *entry, u64 *cookie) argument 368 nfs_readdir_folio_validate(struct folio *folio, u64 last_cookie, u64 change_attr) argument 382 nfs_readdir_folio_unlock_and_put(struct folio *folio) argument 388 nfs_readdir_folio_init_and_validate(struct folio *folio, u64 cookie, u64 change_attr) argument 404 struct folio *folio; local 413 nfs_readdir_folio_last_cookie(struct folio *folio) argument 424 nfs_readdir_folio_needs_filling(struct folio *folio) argument 435 nfs_readdir_folio_set_eof(struct folio *folio) argument 448 struct folio *folio; local 828 struct folio *new, *folio = *arrays; local 934 struct folio *folio = *arrays; local 999 struct folio *folio; local 2608 struct folio *folio; local [all...] |
/linux-master/fs/btrfs/ |
H A D | extent_io.h | 22 struct folio; 70 * folio private set to this value. 119 * For now the folio is always order 0 (aka, a single page). 121 struct folio *folios[INLINE_EXTENT_BUFFER_PAGES]; 155 * 1.1) One large folio covering the whole eb 156 * The eb->start is aligned to folio size, thus adding it 159 * The eb->start is aligned to folio (page) size, thus 163 * In this case there would only be one page sized folio, and there 164 * may be several different extent buffers in the page/folio. 176 * 1.1) One large folio coverin [all...] |
H A D | defrag.c | 863 static struct folio *defrag_prepare_one_folio(struct btrfs_inode *inode, pgoff_t index) 870 struct folio *folio; local 874 folio = __filemap_get_folio(mapping, index, 876 if (IS_ERR(folio)) 877 return folio; 887 if (folio_test_large(folio)) { 888 folio_unlock(folio); 889 folio_put(folio); 893 ret = set_folio_extent_mapped(folio); [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | hugetlb.h | 42 static inline void arch_clear_hugetlb_flags(struct folio *folio) argument 44 clear_bit(PG_arch_1, &folio->flags);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 37 struct folio *last = NULL; 44 struct folio *folio = page_folio(page); local 46 if (folio == last) 48 last = folio; 50 folio_mark_dirty(folio); 52 folio_mark_accessed(folio); 54 if (!folio_batch_add(&fbatch, folio)) 102 struct folio *folio; local [all...] |
/linux-master/include/linux/ |
H A D | bio.h | 256 static inline struct folio *bio_first_folio_all(struct bio *bio) 269 * @folio: The current folio we're iterating. NULL after the last folio. 270 * @offset: The byte offset within the current folio. 271 * @length: The number of bytes in this iteration (will not cross folio 275 struct folio *folio; member in struct:folio_iter 279 struct folio *_next; 290 fi->folio [all...] |
H A D | shmem_fs.h | 144 int shmem_get_folio(struct inode *inode, pgoff_t index, struct folio **foliop, 146 struct folio *shmem_read_folio_gfp(struct address_space *mapping, 149 static inline struct folio *shmem_read_folio(struct address_space *mapping, 193 struct folio **foliop);
|
H A D | gfp.h | 209 struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid, 275 struct folio *__folio_alloc_node_noprof(gfp_t gfp, unsigned int order, int nid) 305 struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order); 306 struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order, struct vm_area_struct *vma, 318 static inline struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order) 336 struct folio *folio = vma_alloc_folio_noprof(gfp, 0, vma, addr, false); local 338 return &folio->page;
|
/linux-master/fs/xfs/scrub/ |
H A D | xfarray.c | 568 * same memory folio. Returns 1 if it sorted, 0 if it did not, or a negative 577 struct folio *folio; local 582 /* No single folio could back this many records. */ 587 folio = xfile_get_folio(si->array->xfile, lo_pos, len, XFILE_ALLOC); 588 if (IS_ERR(folio)) 589 return PTR_ERR(folio); 590 if (!folio) 596 startp = folio_address(folio) + offset_in_folio(folio, lo_po 825 struct folio *folio; local [all...] |
/linux-master/fs/afs/ |
H A D | dir.c | 44 static bool afs_dir_release_folio(struct folio *folio, gfp_t gfp_flags); 45 static void afs_dir_invalidate_folio(struct folio *folio, size_t offset, 49 struct folio *folio) 114 struct folio *folio; local 123 xas_for_each(&xas, folio, last) { 124 if (xas_retry(&xas, folio)) 48 afs_dir_dirty_folio(struct address_space *mapping, struct folio *folio) argument 138 afs_dir_check_folio(struct afs_vnode *dvnode, struct folio *folio, loff_t i_size) argument 188 struct folio *folio; local 222 struct folio *folio; local 322 struct folio *folio; variable in typeref:struct:folio 336 folio_unlock(folio); variable 503 struct folio *folio; local 2027 afs_dir_release_folio(struct folio *folio, gfp_t gfp_flags) argument 2044 afs_dir_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | page.h | 32 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
|
H A D | cacheflush.h | 130 void flush_dcache_folio(struct folio *);
|
/linux-master/arch/arm/mm/ |
H A D | nommu.c | 183 void flush_dcache_folio(struct folio *folio) argument 185 __cpuc_flush_dcache_area(folio_address(folio), folio_size(folio));
|
H A D | fault-armv.c | 188 struct folio *folio; local 200 folio = page_folio(pfn_to_page(pfn)); 201 mapping = folio_flush_mapping(folio); 202 if (!test_and_set_bit(PG_dcache_clean, &folio->flags)) 203 __flush_dcache_folio(mapping, folio);
|
/linux-master/fs/ceph/ |
H A D | addr.c | 67 struct folio **foliop, void **_fsdata); 80 static bool ceph_dirty_folio(struct address_space *mapping, struct folio *folio) argument 87 if (folio_test_dirty(folio)) { 89 ceph_vinop(inode), folio, folio->index); 90 VM_BUG_ON_FOLIO(!folio_test_private(folio), folio); 116 ceph_vinop(inode), folio, folio 137 ceph_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 634 struct folio *folio = page_folio(page); local 1048 struct folio *folio = page_folio(page); local 1493 struct folio *folio = NULL; local 1514 struct folio *folio = page_folio(subpage); local 1823 struct folio *folio = NULL; local [all...] |
/linux-master/kernel/futex/ |
H A D | core.c | 227 struct folio *folio; local 291 * The treatment of mapping from this point on is critical. The folio 292 * lock protects many things but in this context the folio lock 296 * Strictly speaking the folio lock is not needed in all cases being 297 * considered here and folio lock forces unnecessarily serialization. 299 * folio lock will be acquired only if it is unavoidable 301 * Mapping checks require the folio so it is looked up now. For 302 * anonymous pages, it does not matter if the folio is split 307 folio [all...] |
/linux-master/fs/ext4/ |
H A D | mballoc.h | 207 struct folio *ac_bitmap_folio; 208 struct folio *ac_buddy_folio; 218 struct folio *bd_buddy_folio; 220 struct folio *bd_bitmap_folio;
|
/linux-master/arch/riscv/mm/ |
H A D | cacheflush.c | 89 struct folio *folio = page_folio(pte_page(pte)); local 91 if (!test_bit(PG_dcache_clean, &folio->flags)) { 93 set_bit(PG_dcache_clean, &folio->flags);
|
/linux-master/arch/mips/mm/ |
H A D | init.c | 174 struct folio *src = page_folio(from); 200 struct folio *folio = page_folio(page); local 203 folio_mapped(folio) && !folio_test_dcache_dirty(folio)) { 210 folio_set_dcache_dirty(folio); 220 struct folio *folio = page_folio(page); local 223 folio_mapped(folio) && !folio_test_dcache_dirty(folio)) { [all...] |
/linux-master/include/trace/events/ |
H A D | erofs.h | 85 TP_PROTO(struct folio *folio, bool raw), 87 TP_ARGS(folio, raw), 99 __entry->dev = folio->mapping->host->i_sb->s_dev; 100 __entry->nid = EROFS_I(folio->mapping->host)->nid; 101 __entry->dir = S_ISDIR(folio->mapping->host->i_mode); 102 __entry->index = folio->index; 103 __entry->uptodate = folio_test_uptodate(folio);
|
/linux-master/arch/parisc/include/asm/ |
H A D | cacheflush.h | 46 void flush_dcache_folio(struct folio *folio);
|
/linux-master/drivers/misc/lkdtm/ |
H A D | usercopy.c | 405 struct folio *folio; local 412 folio = folio_alloc(GFP_KERNEL | __GFP_ZERO, 1); 413 if (!folio) { 417 addr = folio_address(folio); 419 do_usercopy_page_span("folio", addr + PAGE_SIZE); 422 folio_put(folio);
|