/linux-master/mm/ |
H A D | slub.c | 961 struct folio *folio = (struct folio *)slab_folio(slab); local 965 folio_flags(folio, 0)); 2176 struct folio *folio; local 2180 folio = (struct folio *)alloc_pages_node(node, flags, order); 2181 if (!folio) 2184 slab = folio_slab(folio); 2402 struct folio *folio = slab_folio(slab); local 3909 struct folio *folio; local 4354 free_large_kmalloc(struct folio *folio, void *object) argument 4378 struct folio *folio; local 4426 struct folio *folio; local [all...] |
H A D | zswap.c | 790 void zswap_folio_swapin(struct folio *folio) argument 794 if (folio) { 795 lruvec = folio_lruvec(folio); 1000 static bool zswap_compress(struct folio *folio, struct zswap_entry *entry) argument 1018 sg_set_page(&input, &folio->page, PAGE_SIZE, 0); 1115 * Attempts to free an entry by adding a folio to the swap cache, 1116 * decompressing the entry data into the folio, and issuing a 1117 * bio write to write the folio bac 1130 struct folio *folio; local 1498 zswap_store(struct folio *folio) argument 1641 zswap_load(struct folio *folio) argument [all...] |
H A D | hugetlb.c | 59 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument 61 return cma_pages_valid(hugetlb_cma[folio_nid(folio)], &folio->page, 65 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument 1311 static void enqueue_hugetlb_folio(struct hstate *h, struct folio *folio) argument 1313 int nid = folio_nid(folio); 1316 VM_BUG_ON_FOLIO(folio_ref_count(folio), foli local 1327 struct folio *folio; local 1363 struct folio *folio; local 1395 struct folio *folio = NULL; local 1512 __destroy_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument 1535 destroy_compound_hugetlb_folio_for_demote(struct folio *folio, unsigned int order) argument 1542 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 1548 free_gigantic_folio(struct folio *folio, unsigned int order) argument 1616 free_gigantic_folio(struct folio *folio, unsigned int order) argument 1618 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 1622 __clear_hugetlb_destructor(struct hstate *h, struct folio *folio) argument 1640 __remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus, bool demote) argument 1646 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio(folio), folio); local 1647 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio_rsvd(folio), folio); local 1683 remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1689 remove_hugetlb_folio_for_demote(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1695 add_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1741 __update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument 1823 struct folio *folio; local 1851 update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio, bool atomic) argument 1874 struct folio *folio, *t_folio; local 1925 struct folio *folio, *t_folio; local 1973 free_huge_folio(struct folio *folio) argument 1985 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local 1986 VM_BUG_ON_FOLIO(folio_mapcount(folio), folio); local 2017 pages_per_huge_page(h), folio); local 2019 pages_per_huge_page(h), folio); local 2050 init_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument 2059 __prep_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument 2065 prep_new_hugetlb_folio(struct hstate *h, struct folio *folio, int nid) argument 2073 __prep_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument 2150 prep_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 2156 prep_compound_gigantic_folio_for_demote(struct folio *folio, unsigned int order) argument 2249 struct folio *folio; local 2283 struct folio *folio; local 2303 struct folio *folio; local 2318 struct folio *folio, *tmp_f; local 2345 struct folio *folio; local 2367 struct folio *folio = NULL; local 2404 struct folio *folio = page_folio(page); local 2515 struct folio *folio = NULL; local 2556 struct folio *folio; local 2583 struct folio *folio = NULL; local 2612 struct folio *folio; local 2634 struct folio *folio, *tmp; variable in typeref:struct:folio 2750 struct folio *folio; local 2932 restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma, unsigned long address, struct folio *folio) argument 3099 struct folio *folio = page_folio(page); local 3137 struct folio *folio; local 3258 pages_per_huge_page(h), folio); local 3335 hugetlb_folio_init_tail_vmemmap(struct folio *folio, unsigned long start_page_number, unsigned long end_page_number) argument 3355 hugetlb_folio_init_vmemmap(struct folio *folio, struct hstate *h, unsigned long nr_pages) argument 3375 struct folio *folio, *tmp_f; local 3412 struct folio *folio = (void *)page; local 3478 struct folio *folio; local 3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], local 3706 struct folio *folio, *next; local 3769 struct folio *folio; local 3936 demote_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument 4013 struct folio *folio; variable in typeref:struct:folio 6124 struct folio *folio; local 6133 hugetlb_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t idx) argument 6207 struct folio *folio; local 6439 struct folio *folio = NULL; local 6640 struct folio *folio; local 6674 struct folio *folio; local 7592 isolate_hugetlb(struct folio *folio, struct list_head *list) argument 7610 get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison) argument 7640 folio_putback_active_hugetlb(struct folio *folio) argument [all...] |
H A D | shmem.c | 150 struct folio **foliop, enum sgp_type sgp, gfp_t gfp, 625 struct folio *folio; local 678 folio = filemap_get_folio(inode->i_mapping, index); 679 if (IS_ERR(folio)) 683 if (!folio_test_large(folio)) { 684 folio_put(folio); 695 if (!folio_trylock(folio)) { 696 folio_put(folio); 700 ret = split_folio(folio); 761 shmem_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t index, void *expected, gfp_t gfp) argument 814 shmem_delete_from_page_cache(struct folio *folio, void *radswap) argument 934 struct folio *folio; local 973 struct folio *folio; local 1083 folio); local 1278 struct folio *folio; local 1324 struct folio *folio = fbatch->folios[i]; local 1421 struct folio *folio = page_folio(page); local 1573 struct folio *folio; local 1640 struct folio *folio; local 1743 shmem_should_replace_folio(struct folio *folio, gfp_t gfp) argument 1817 shmem_set_folio_swapin_error(struct inode *inode, pgoff_t index, struct folio *folio, swp_entry_t swap) argument 1856 struct folio *folio = NULL; local 1965 struct folio *folio; local 2240 struct folio *folio = NULL; local 2618 struct folio *folio; local 2679 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local 2732 struct folio *folio; local 2764 struct folio *folio = page_folio(page); local 2800 struct folio *folio = NULL; local 2977 struct folio *folio = NULL; local 3167 struct folio *folio; local 3500 struct folio *folio; local 3564 struct folio *folio = NULL; local 4510 shmem_error_remove_folio(struct address_space *mapping, struct folio *folio) argument 4957 struct folio *folio; local 4979 struct folio *folio = shmem_read_folio_gfp(mapping, index, gfp); local [all...] |
H A D | page_owner.c | 367 void __folio_copy_owner(struct folio *newfolio, struct folio *old) 395 * Do not proactively clear PAGE_EXT_OWNER{_ALLOCATED} bits as the folio 404 * We linked the original stack to the new folio, we need to do the same 405 * for the new one and the old folio otherwise there will be an imbalance
|
H A D | memory-failure.c | 610 static void collect_procs_anon(struct folio *folio, struct page *page, argument 618 av = folio_lock_anon_vma_read(folio, NULL); 647 static void collect_procs_file(struct folio *folio, struct page *page, argument 652 struct address_space *mapping = folio->mapping; 725 static void collect_procs(struct folio *folio, struct page *page, argument 728 if (!folio->mapping) 733 collect_procs_anon(folio, pag 923 delete_from_lru_cache(struct folio *folio) argument 948 truncate_error_folio(struct folio *folio, unsigned long pfn, struct address_space *mapping) argument 1034 struct folio *folio = page_folio(p); local 1158 struct folio *folio = page_folio(p); local 1180 struct folio *folio = page_folio(p); local 1202 struct folio *folio = page_folio(p); local 1404 struct folio *folio = page_folio(page); local 1501 struct folio *folio = page_folio(page); local 1573 struct folio *folio = page_folio(hpage); local 1748 struct folio *folio = pfn_folio(pfn); local 1859 raw_hwp_list_head(struct folio *folio) argument 1868 struct folio *folio = page_folio(page); local 1899 __folio_free_raw_hwp(struct folio *folio, bool move_flag) argument 1917 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument 1960 folio_free_raw_hwp(struct folio *folio, bool move_flag) argument 1979 folio_clear_hugetlb_hwpoison(struct folio *folio) argument 2003 struct folio *folio = page_folio(page); local 2056 struct folio *folio; local 2125 folio_free_raw_hwp(struct folio *folio, bool flag) argument 2530 struct folio *folio; local 2631 mf_isolate_folio(struct folio *folio, struct list_head *pagelist) argument 2674 struct folio *folio = page_folio(page); local [all...] |
H A D | madvise.c | 183 struct folio *folio; local 201 folio = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE, 203 if (folio) 204 folio_put(folio); 226 struct folio *folio; local 230 xas_for_each(&xas, folio, end_index) { 234 if (!xa_is_value(folio)) 236 entry = radix_to_swp_entry(folio); 335 struct folio *folio = NULL; local 491 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local 628 struct folio *folio; local [all...] |
H A D | internal.h | 56 * If a 16GB hugetlb folio were mapped by PTEs of all of its 4kB pages, 72 * the folio's entire_mapcount. 74 static inline int folio_nr_pages_mapped(struct folio *folio) argument 76 return atomic_read(&folio->_nr_pages_mapped) & FOLIO_PAGES_MAPPED; 79 static inline void *folio_raw_mapping(struct folio *folio) argument 81 unsigned long mapping = (unsigned long)folio->mapping; 107 * folio_pte_batch - detect a PTE batch for a large folio 108 * @folio 129 folio_pte_batch(struct folio *folio, unsigned long addr, pte_t *start_ptep, pte_t pte, int max_nr, fpb_t flags, bool *any_writable) argument 144 VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio); local 181 acct_reclaim_writeback(struct folio *folio) argument 250 folio_evictable(struct folio *folio) argument 276 folio_needs_release(struct folio *folio) argument 499 folio_set_order(struct folio *folio, unsigned int order) argument 514 struct folio *folio = (struct folio *)page; local 522 struct folio *folio = (struct folio *)page; local 710 folio_within_range(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 716 VM_WARN_ON_FOLIO(folio_test_ksm(folio), folio); local 738 folio_within_vma(struct folio *folio, struct vm_area_struct *vma) argument 753 mlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 769 munlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 872 unmap_mapping_folio(struct folio *folio) argument 873 mlock_new_folio(struct folio *folio) argument [all...] |
H A D | huge_memory.c | 771 struct deferred_split *get_deferred_split_queue(struct folio *folio) argument 773 struct mem_cgroup *memcg = folio_memcg(folio); 774 struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); 783 struct deferred_split *get_deferred_split_queue(struct folio *folio) argument 785 struct pglist_data *pgdat = NODE_DATA(folio_nid(folio)); 791 void folio_prep_large_rmappable(struct folio *folio) argument 793 if (!folio || !folio_test_larg 800 is_transparent_hugepage(struct folio *folio) argument 872 struct folio *folio = page_folio(page); local 1000 struct folio *folio; local 1517 struct folio *folio; local 1718 struct folio *folio; local 1807 struct folio *folio; local 1910 struct folio *folio = NULL; local 2053 struct folio *folio = pfn_swap_entry_folio(entry); local 2084 struct folio *folio; local 2436 struct folio *folio; local 2649 __split_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd, unsigned long address, bool freeze, struct folio *folio) argument 2684 split_huge_pmd_address(struct vm_area_struct *vma, unsigned long address, bool freeze, struct folio *folio) argument 2730 unmap_folio(struct folio *folio) argument 2753 remap_page(struct folio *folio, unsigned long nr) argument 2792 __split_huge_page_tail(struct folio *folio, int tail, struct lruvec *lruvec, struct list_head *list, unsigned int new_order) argument 2892 struct folio *folio = page_folio(page); local 3000 can_split_folio(struct folio *folio, int *pextra_pins) argument 3042 struct folio *folio = page_folio(page); local 3230 folio_undo_large_rmappable(struct folio *folio) argument 3255 deferred_split_folio(struct folio *folio) argument 3320 struct folio *folio, *next; local 3373 struct folio *folio; local 3468 struct folio *folio; local 3547 struct folio *folio = filemap_get_folio(mapping, index); local 3676 struct folio *folio = page_folio(page); local 3725 struct folio *folio = page_folio(new); local [all...] |
H A D | gup.c | 53 struct folio *folio = page_folio(page); local 56 !folio_test_anon(folio)) 58 if (!folio_test_large(folio) || folio_test_hugetlb(folio)) 59 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page), page); 62 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page) && 68 * Return the folio with ref appropriately incremented, 71 static inline struct folio *try_get_folio(struct page *page, int refs) 73 struct folio *foli local 128 struct folio *folio; local 189 gup_put_folio(struct folio *folio, int refs, unsigned int flags) argument 227 struct folio *folio = page_folio(page); local 286 folio_add_pin(struct folio *folio) argument 310 struct folio *folio = page_folio(next); local 324 struct folio *folio = page_folio(list[i]); local 362 struct folio *folio; local 428 struct folio *folio; local 446 struct folio *folio; local 472 struct folio *folio; local 1307 struct folio *folio; local 2063 struct folio *folio = page_folio(pages[i]); local 2113 struct folio *folio = page_folio(pages[i]); local 2484 folio_fast_pin_allowed(struct folio *folio, unsigned int flags) argument 2593 struct folio *folio; local 2807 struct folio *folio; local 2881 struct folio *folio; local 2925 struct folio *folio; local 2971 struct folio *folio; local [all...] |
/linux-master/lib/ |
H A D | scatterlist.c | 1276 struct folio *folio; local 1285 xas_for_each(&xas, folio, ULONG_MAX) { 1286 if (xas_retry(&xas, folio)) 1288 if (WARN_ON(xa_is_value(folio))) 1290 if (WARN_ON(folio_test_hugetlb(folio))) 1293 offset = offset_in_folio(folio, start); 1294 len = min_t(size_t, maxsize, folio_size(folio) - offset); 1296 sg_set_page(sg, folio_page(folio, 0), len, offset);
|
/linux-master/arch/xtensa/include/asm/ |
H A D | cacheflush.h | 127 void flush_dcache_folio(struct folio *folio);
|
/linux-master/kernel/sched/ |
H A D | fair.c | 1770 static int numa_hint_fault_latency(struct folio *folio) argument 1775 last_time = folio_xchg_access_time(folio, time); 1832 bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio, argument 1868 latency = numa_hint_fault_latency(folio); 1873 folio_nr_pages(folio)); 1877 last_cpupid = folio_xchg_last_cpupid(folio, this_cpupid);
|
/linux-master/fs/erofs/ |
H A D | fscache.c | 57 struct folio *folio; local 65 xas_for_each(&xas, folio, last_page) { 66 if (xas_retry(&xas, folio)) 69 folio_mark_uptodate(folio); 70 folio_unlock(folio); 224 static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio) argument 226 struct erofs_fscache *ctx = folio->mapping->host->i_private; 231 req = erofs_fscache_req_alloc(folio 345 erofs_fscache_read_folio(struct file *file, struct folio *folio) argument [all...] |
/linux-master/include/linux/ |
H A D | page-flags.h | 144 * Depending on the way an anonymous folio can be mapped into a page 147 * tail pages of an anonymous folio. For now, we only expect it to be 191 /* At least one page in this folio has the hwpoison flag set */ 256 * page_folio - Converts from page to folio. 259 * Every page is part of a folio. This function cannot be called on a 263 * does not hold a reference, this call may race with a folio split, so 264 * it should re-check the folio still contains this page after gaining 265 * a reference on the folio. 266 * Return: The folio which contains this page. 269 const struct page *: (const struct folio *)_compound_hea 308 const_folio_flags(const struct folio *folio, unsigned n) argument 318 folio_flags(struct folio *folio, unsigned n) argument 581 folio_test_swapcache(const struct folio *folio) argument 686 folio_mapping_flags(const struct folio *folio) argument 696 folio_test_anon(const struct folio *folio) argument 706 __folio_test_movable(const struct folio *folio) argument 725 folio_test_ksm(const struct folio *folio) argument 754 folio_xor_flags_has_waiters(struct folio *folio, unsigned long mask) argument 770 folio_test_uptodate(const struct folio *folio) argument 792 __folio_mark_uptodate(struct folio *folio) argument 798 folio_mark_uptodate(struct folio *folio) argument 829 folio_test_head(const struct folio *folio) argument 850 folio_test_large(const struct folio *folio) argument 1157 folio_has_private(const struct folio *folio) argument [all...] |
H A D | mm.h | 215 #define folio_page_idx(folio, p) (page_to_pfn(p) - folio_pfn(folio)) 218 #define folio_page_idx(folio, p) ((p) - &(folio)->page) 230 static inline struct folio *lru_to_folio(struct list_head *head) 232 return list_entry((head)->prev, struct folio, lru); 1091 struct folio *folio = (struct folio *)page; local 1093 if (!test_bit(PG_head, &folio 1107 folio_order(struct folio *folio) argument 1138 folio_put_testzero(struct folio *folio) argument 1202 folio_entire_mapcount(struct folio *folio) argument 1255 folio_mapcount(struct folio *folio) argument 1262 folio_large_is_mapped(struct folio *folio) argument 1278 folio_mapped(struct folio *folio) argument 1472 folio_get(struct folio *folio) argument 1474 VM_BUG_ON_FOLIO(folio_ref_zero_or_close_to_overflow(folio), folio); local 1505 folio_put(struct folio *folio) argument 1525 folio_put_refs(struct folio *folio, int refs) argument 1572 struct folio *folio = page_folio(page); local 1668 folio_nid(const struct folio *folio) argument 1723 folio_xchg_last_cpupid(struct folio *folio, int cpupid) argument 1728 folio_last_cpupid(struct folio *folio) argument 1737 folio_last_cpupid(struct folio *folio) argument 1750 folio_xchg_access_time(struct folio *folio, int time) argument 1769 folio_xchg_last_cpupid(struct folio *folio, int cpupid) argument 1774 folio_xchg_access_time(struct folio *folio, int time) argument 1779 folio_last_cpupid(struct folio *folio) argument 1887 folio_zone(const struct folio *folio) argument 1892 folio_pgdat(const struct folio *folio) argument 1919 folio_pfn(struct folio *folio) argument 1954 folio_maybe_dma_pinned(struct folio *folio) argument 1982 folio_needs_cow_for_dma(struct vm_area_struct *vma, struct folio *folio) argument 2010 is_zero_folio(const struct folio *folio) argument 2017 folio_is_longterm_pinnable(struct folio *folio) argument 2038 folio_is_longterm_pinnable(struct folio *folio) argument 2072 folio_nr_pages(struct folio *folio) argument 2097 struct folio *folio = (struct folio *)page; local 2131 folio_next(struct folio *folio) argument 2148 folio_shift(struct folio *folio) argument 2161 folio_size(struct folio *folio) argument 2179 folio_estimated_sharers(struct folio *folio) argument 2192 arch_make_folio_accessible(struct folio *folio) argument 2245 folio_address(const struct folio *folio) argument 2283 folio_is_pfmemalloc(const struct folio *folio) argument 2621 mm_counter_file(struct folio *folio) argument 2628 mm_counter(struct folio *folio) argument 2949 struct folio *folio = ptdesc_folio(ptdesc); local 2960 struct folio *folio = ptdesc_folio(ptdesc); local 3064 struct folio *folio = ptdesc_folio(ptdesc); local 3075 struct folio *folio = ptdesc_folio(ptdesc); local 3103 struct folio *folio = ptdesc_folio(ptdesc); local 3111 struct folio *folio = ptdesc_folio(ptdesc); local [all...] |
H A D | swapops.h | 500 static inline struct folio *pfn_swap_entry_folio(swp_entry_t entry) 502 struct folio *folio = pfn_folio(swp_offset_pfn(entry)); local 506 * corresponding folio is locked 508 BUG_ON(is_migration_entry(entry) && !folio_test_locked(folio)); 510 return folio;
|
H A D | shmem_fs.h | 144 int shmem_get_folio(struct inode *inode, pgoff_t index, struct folio **foliop, 146 struct folio *shmem_read_folio_gfp(struct address_space *mapping, 149 static inline struct folio *shmem_read_folio(struct address_space *mapping, 193 struct folio **foliop);
|
/linux-master/fs/ntfs3/ |
H A D | inode.c | 565 struct folio *folio = bh->b_folio; local 580 err = attr_data_read_resident(ni, &folio->page); 653 if (folio) { 658 folio_set_bh(bh, folio, off); 664 folio_zero_segment(folio, off + voff, off + block_size); 705 static int ntfs_read_folio(struct file *file, struct folio *folio) argument 707 struct page *page = &folio->page; 731 return mpage_read_folio(folio, ntfs_get_bloc 857 ntfs_resident_writepage(struct folio *folio, struct writeback_control *wbc, void *data) argument [all...] |
H A D | file.c | 190 struct folio *folio; local 199 folio = __filemap_get_folio(mapping, idx, 202 if (IS_ERR(folio)) 203 return PTR_ERR(folio); 205 head = folio_buffers(folio); 207 head = create_empty_buffers(folio, blocksize, 0); 225 if (folio_test_uptodate(folio)) 229 folio_unlock(folio); 230 folio_put(folio); [all...] |
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 21 NETFS_JUST_PREFETCH, /* We have to read the folio anyway */ 22 NETFS_WHOLE_FOLIO_MODIFY, /* We're going to overwrite the whole folio */ 31 static void netfs_set_group(struct folio *folio, struct netfs_group *netfs_group) argument 33 if (netfs_group && !folio_get_private(folio)) 34 folio_attach_private(folio, netfs_get_group(netfs_group)); 38 static void netfs_folio_start_fscache(bool caching, struct folio *folio) argument 41 folio_start_fscache(folio); 44 static void netfs_folio_start_fscache(bool caching, struct folio *foli argument 56 netfs_how_to_modify(struct netfs_inode *ctx, struct file *file, struct folio *folio, void *netfs_group, size_t flen, size_t offset, size_t len, bool maybe_trouble) argument 163 struct folio *folio; local 508 struct folio *folio = page_folio(vmf->page); local 567 struct folio *folio; local 605 struct folio *folio; local 642 struct folio *folio; local 781 struct folio *folio; local 892 netfs_write_back_from_locked_folio(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, struct folio *folio, unsigned long long start, unsigned long long end) argument 1006 struct folio *folio; local 1204 netfs_launder_folio(struct folio *folio) argument [all...] |
/linux-master/arch/arc/mm/ |
H A D | tlb.c | 490 struct folio *folio = page_folio(page); local 491 int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags); 493 unsigned long offset = offset_in_folio(folio, paddr); 494 nr = folio_nr_pages(folio);
|
/linux-master/fs/btrfs/ |
H A D | inode.c | 4740 struct folio *folio; local 4772 folio = __filemap_get_folio(mapping, index, 4774 if (IS_ERR(folio)) { 4782 if (!folio_test_uptodate(folio)) { 4783 ret = btrfs_read_folio(NULL, folio); 4784 folio_lock(folio); 4785 if (folio->mapping != mapping) { 4786 folio_unlock(folio); 4787 folio_put(folio); 7947 struct folio *folio = page_folio(page); local 7971 __btrfs_release_folio(struct folio *folio, gfp_t gfp_flags) argument 7982 btrfs_release_folio(struct folio *folio, gfp_t gfp_flags) argument 8010 btrfs_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 8195 struct folio *folio = page_folio(page); local [all...] |
/linux-master/fs/smb/client/ |
H A D | smb2ops.c | 4311 struct folio *folio; local 4316 xas_for_each_marked(&xas, folio, ULONG_MAX, XA_MARK_0) { 4317 folio_put(folio);
|
H A D | cifsfs.c | 1237 * Flush out either the folio that overlaps the beginning of a range in which 1238 * pos resides or the folio that overlaps the end of a range unless that folio 1240 * bounds to encompass the folio. 1245 struct folio *folio; local 1251 folio = filemap_get_folio(inode->i_mapping, index); 1252 if (IS_ERR(folio)) 1255 size = folio_size(folio); 1256 fpos = folio_pos(folio); [all...] |