/linux-master/mm/damon/ |
H A D | vaddr.c | 343 struct folio *folio = pfn_folio(pte_pfn(entry)); local 346 folio_get(folio); 361 folio_set_young(folio); 363 folio_set_idle(folio); 364 folio_put(folio); 432 /* size of the folio for the access checked virtual memory address */ 443 struct folio *folio; local 462 folio 507 struct folio *folio; local [all...] |
/linux-master/fs/erofs/ |
H A D | fscache.c | 57 struct folio *folio; local 65 xas_for_each(&xas, folio, last_page) { 66 if (xas_retry(&xas, folio)) 69 folio_mark_uptodate(folio); 70 folio_unlock(folio); 224 static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio) argument 226 struct erofs_fscache *ctx = folio->mapping->host->i_private; 231 req = erofs_fscache_req_alloc(folio 345 erofs_fscache_read_folio(struct file *file, struct folio *folio) argument [all...] |
/linux-master/fs/jffs2/ |
H A D | file.c | 30 static int jffs2_read_folio(struct file *filp, struct folio *folio); 113 int __jffs2_read_folio(struct file *file, struct folio *folio) argument 115 int ret = jffs2_do_readpage_nolock(folio->mapping->host, &folio->page); 116 folio_unlock(folio); 120 static int jffs2_read_folio(struct file *file, struct folio *folio) argument 122 struct jffs2_inode_info *f = JFFS2_INODE_INFO(folio [all...] |
/linux-master/fs/hugetlbfs/ |
H A D | inode.c | 325 struct folio *folio; local 343 /* Find the folio */ 344 folio = filemap_lock_hugetlb_folio(h, mapping, index); 345 if (IS_ERR(folio)) { 352 folio_unlock(folio); 354 if (!folio_test_hwpoison(folio)) 362 want = adjust_range_hwpoison(&folio->page, offset, nr); 364 folio_put(folio); 371 * We have the folio, cop 406 hugetlb_delete_from_page_cache(struct folio *folio) argument 471 hugetlb_unmap_file_folio(struct hstate *h, struct address_space *mapping, struct folio *folio, pgoff_t index) argument 590 remove_inode_single_folio(struct hstate *h, struct inode *inode, struct address_space *mapping, struct folio *folio, pgoff_t index, bool truncate_op) argument 614 VM_BUG_ON_FOLIO(folio_test_hugetlb_restore_reserve(folio), folio); local 662 struct folio *folio = fbatch.folios[i]; local 731 struct folio *folio; local 855 struct folio *folio; local 1145 hugetlbfs_error_remove_folio(struct address_space *mapping, struct folio *folio) argument [all...] |
/linux-master/arch/arm/mm/ |
H A D | mm.h | 48 void __flush_dcache_folio(struct address_space *mapping, struct folio *folio);
|
H A D | nommu.c | 183 void flush_dcache_folio(struct folio *folio) argument 185 __cpuc_flush_dcache_area(folio_address(folio), folio_size(folio));
|
/linux-master/arch/arc/include/asm/ |
H A D | cacheflush.h | 31 void flush_dcache_folio(struct folio *folio);
|
/linux-master/arch/nios2/include/asm/ |
H A D | cacheflush.h | 32 void flush_dcache_folio(struct folio *folio);
|
/linux-master/fs/netfs/ |
H A D | iterator.c | 154 struct folio *folio; local 169 xas_for_each(&xas, folio, ULONG_MAX) { 171 if (xas_retry(&xas, folio)) 173 if (WARN_ON(xa_is_value(folio))) 175 if (WARN_ON(folio_test_hugetlb(folio))) 178 flen = folio_size(folio); 179 offset = offset_in_folio(folio, pos);
|
/linux-master/fs/btrfs/ |
H A D | extent_io.h | 22 struct folio; 69 * folio private set to this value. 118 * For now the folio is always order 0 (aka, a single page). 120 struct folio *folios[INLINE_EXTENT_BUFFER_PAGES]; 154 * 1.1) One large folio covering the whole eb 155 * The eb->start is aligned to folio size, thus adding it 158 * The eb->start is aligned to folio (page) size, thus 162 * In this case there would only be one page sized folio, and there 163 * may be several different extent buffers in the page/folio. 175 * 1.1) One large folio coverin [all...] |
H A D | bio.h | 110 u64 length, u64 logical, struct folio *folio,
|
/linux-master/mm/ |
H A D | slab.h | 110 * folio_slab - Converts from folio to slab. 111 * @folio: The folio. 113 * Currently struct slab is a different representation of a folio where 116 * Return: The slab which contains this folio. 118 #define folio_slab(folio) (_Generic((folio), \ 119 const struct folio *: (const struct slab *)(folio), \ 120 struct folio * 204 struct folio *folio = virt_to_folio(addr); local [all...] |
H A D | hugetlb_cgroup.c | 194 struct folio *folio = page_folio(page); local 196 page_hcg = hugetlb_cgroup_from_folio(folio); 215 set_hugetlb_cgroup(folio, parent); 307 struct folio *folio, bool rsvd) 312 __set_hugetlb_cgroup(folio, h_cg, rsvd); 315 h_cg->nodeinfo[folio_nid(folio)]->usage[idx]; 321 WRITE_ONCE(h_cg->nodeinfo[folio_nid(folio)]->usage[idx], 328 struct folio *foli 305 __hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio, bool rsvd) argument 326 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 333 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 343 __hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio, bool rsvd) argument 375 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument 381 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument [all...] |
H A D | zswap.c | 790 void zswap_folio_swapin(struct folio *folio) argument 794 if (folio) { 795 lruvec = folio_lruvec(folio); 1000 static bool zswap_compress(struct folio *folio, struct zswap_entry *entry) argument 1018 sg_set_page(&input, &folio->page, PAGE_SIZE, 0); 1115 * Attempts to free an entry by adding a folio to the swap cache, 1116 * decompressing the entry data into the folio, and issuing a 1117 * bio write to write the folio bac 1130 struct folio *folio; local 1498 zswap_store(struct folio *folio) argument 1641 zswap_load(struct folio *folio) argument [all...] |
/linux-master/fs/affs/ |
H A D | file.c | 380 static int affs_read_folio(struct file *file, struct folio *folio) argument 382 return block_read_full_folio(folio, affs_get_block); 526 static int affs_do_read_folio_ofs(struct folio *folio, size_t to, int create) argument 528 struct inode *inode = folio->mapping->host; 536 folio->index, to); 537 BUG_ON(to > folio_size(folio)); 539 tmp = folio_pos(folio); 549 memcpy_to_folio(folio, po 629 affs_read_folio_ofs(struct file *file, struct folio *folio) argument 654 struct folio *folio; local 692 struct folio *folio = page_folio(page); local [all...] |
/linux-master/fs/nfs/ |
H A D | dir.c | 58 static void nfs_readdir_clear_array(struct folio *); 157 struct folio *folio; member in struct:nfs_readdir_descriptor 201 static void nfs_readdir_folio_init_array(struct folio *folio, u64 last_cookie, argument 206 array = kmap_local_folio(folio, 0); 219 static void nfs_readdir_clear_array(struct folio *folio) argument 224 array = kmap_local_folio(folio, 0); 231 static void nfs_readdir_folio_reinit_array(struct folio *foli argument 241 struct folio *folio = folio_alloc(gfp_flags, 0); local 247 nfs_readdir_folio_array_free(struct folio *folio) argument 309 nfs_readdir_folio_array_append(struct folio *folio, const struct nfs_entry *entry, u64 *cookie) argument 366 nfs_readdir_folio_validate(struct folio *folio, u64 last_cookie, u64 change_attr) argument 380 nfs_readdir_folio_unlock_and_put(struct folio *folio) argument 386 nfs_readdir_folio_init_and_validate(struct folio *folio, u64 cookie, u64 change_attr) argument 402 struct folio *folio; local 411 nfs_readdir_folio_last_cookie(struct folio *folio) argument 422 nfs_readdir_folio_needs_filling(struct folio *folio) argument 433 nfs_readdir_folio_set_eof(struct folio *folio) argument 446 struct folio *folio; local 826 struct folio *new, *folio = *arrays; local 932 struct folio *folio = *arrays; local 997 struct folio *folio; local 2537 struct folio *folio; local [all...] |
/linux-master/arch/parisc/kernel/ |
H A D | cache.c | 100 struct folio *folio; local 109 folio = page_folio(pfn_to_page(pfn)); 110 pfn = folio_pfn(folio); 111 nr = folio_nr_pages(folio); 112 if (folio_flush_mapping(folio) && 113 test_bit(PG_dcache_dirty, &folio->flags)) { 116 clear_bit(PG_dcache_dirty, &folio->flags); 413 void flush_dcache_folio(struct folio *folio) argument [all...] |
/linux-master/arch/sh/mm/ |
H A D | cache-sh4.c | 112 struct folio *folio = arg; local 114 struct address_space *mapping = folio_flush_mapping(folio); 117 clear_bit(PG_dcache_clean, &folio->flags); 121 unsigned long pfn = folio_pfn(folio); 122 unsigned long addr = (unsigned long)folio_address(folio); 123 unsigned int i, nr = folio_nr_pages(folio);
|
/linux-master/fs/udf/ |
H A D | symlink.c | 100 static int udf_symlink_filler(struct file *file, struct folio *folio) argument 102 struct page *page = &folio->page;
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 37 struct folio *last = NULL; 44 struct folio *folio = page_folio(page); local 46 if (folio == last) 48 last = folio; 50 folio_mark_dirty(folio); 52 folio_mark_accessed(folio); 54 if (!folio_batch_add(&fbatch, folio)) 102 struct folio *folio; local [all...] |
/linux-master/include/linux/ |
H A D | bio.h | 256 static inline struct folio *bio_first_folio_all(struct bio *bio) 269 * @folio: The current folio we're iterating. NULL after the last folio. 270 * @offset: The byte offset within the current folio. 271 * @length: The number of bytes in this iteration (will not cross folio 275 struct folio *folio; member in struct:folio_iter 279 struct folio *_next; 290 fi->folio [all...] |
/linux-master/fs/smb/client/ |
H A D | file.c | 46 struct folio *folio; local 54 xas_for_each_marked(&xas, folio, end, PAGECACHE_TAG_DIRTY) { 55 if (xas_retry(&xas, folio)) 59 folio_lock(folio); 60 folio_clear_dirty_for_io(folio); 61 folio_unlock(folio); 74 struct folio *folio; local 85 xas_for_each(&xas, folio, en 107 struct folio *folio; local 140 struct folio *folio; local 2729 struct folio *folio; local 2830 cifs_write_back_from_locked_folio(struct address_space *mapping, struct writeback_control *wbc, struct xa_state *xas, struct folio *folio, unsigned long long start, unsigned long long end) argument 2976 struct folio *folio; local 3178 struct folio *folio = page_folio(page); local 4507 struct folio *folio = page_folio(vmf->page); local 4573 struct folio *folio; local 4587 struct folio *folio; local 4655 struct folio *folio; local 4828 cifs_read_folio(struct file *file, struct folio *folio) argument 4976 cifs_release_folio(struct folio *folio, gfp_t gfp) argument 4989 cifs_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 4995 cifs_launder_folio(struct folio *folio) argument [all...] |
/linux-master/fs/afs/ |
H A D | dir.c | 44 static bool afs_dir_release_folio(struct folio *folio, gfp_t gfp_flags); 45 static void afs_dir_invalidate_folio(struct folio *folio, size_t offset, 49 struct folio *folio) 114 struct folio *folio; local 123 xas_for_each(&xas, folio, last) { 124 if (xas_retry(&xas, folio)) 48 afs_dir_dirty_folio(struct address_space *mapping, struct folio *folio) argument 138 afs_dir_check_folio(struct afs_vnode *dvnode, struct folio *folio, loff_t i_size) argument 188 struct folio *folio; local 222 struct folio *folio; local 322 struct folio *folio; variable in typeref:struct:folio 336 folio_unlock(folio); variable 503 struct folio *folio; local 2027 afs_dir_release_folio(struct folio *folio, gfp_t gfp_flags) argument 2044 afs_dir_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | page.h | 32 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
|
H A D | cacheflush.h | 130 void flush_dcache_folio(struct folio *);
|