Searched refs:folio (Results 1 - 25 of 361) sorted by last modified time

1234567891011>>

/linux-master/mm/
H A Dslub.c961 struct folio *folio = (struct folio *)slab_folio(slab); local
965 folio_flags(folio, 0));
2176 struct folio *folio; local
2180 folio = (struct folio *)alloc_pages_node(node, flags, order);
2181 if (!folio)
2184 slab = folio_slab(folio);
2402 struct folio *folio = slab_folio(slab); local
3909 struct folio *folio; local
4354 free_large_kmalloc(struct folio *folio, void *object) argument
4378 struct folio *folio; local
4426 struct folio *folio; local
[all...]
H A Dzswap.c790 void zswap_folio_swapin(struct folio *folio) argument
794 if (folio) {
795 lruvec = folio_lruvec(folio);
1000 static bool zswap_compress(struct folio *folio, struct zswap_entry *entry) argument
1018 sg_set_page(&input, &folio->page, PAGE_SIZE, 0);
1115 * Attempts to free an entry by adding a folio to the swap cache,
1116 * decompressing the entry data into the folio, and issuing a
1117 * bio write to write the folio bac
1130 struct folio *folio; local
1498 zswap_store(struct folio *folio) argument
1641 zswap_load(struct folio *folio) argument
[all...]
H A Dhugetlb.c59 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument
61 return cma_pages_valid(hugetlb_cma[folio_nid(folio)], &folio->page,
65 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument
1311 static void enqueue_hugetlb_folio(struct hstate *h, struct folio *folio) argument
1313 int nid = folio_nid(folio);
1316 VM_BUG_ON_FOLIO(folio_ref_count(folio), foli local
1327 struct folio *folio; local
1363 struct folio *folio; local
1395 struct folio *folio = NULL; local
1512 __destroy_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument
1535 destroy_compound_hugetlb_folio_for_demote(struct folio *folio, unsigned int order) argument
1542 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
1548 free_gigantic_folio(struct folio *folio, unsigned int order) argument
1616 free_gigantic_folio(struct folio *folio, unsigned int order) argument
1618 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
1622 __clear_hugetlb_destructor(struct hstate *h, struct folio *folio) argument
1640 __remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus, bool demote) argument
1646 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio(folio), folio); local
1647 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio_rsvd(folio), folio); local
1683 remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1689 remove_hugetlb_folio_for_demote(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1695 add_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument
1741 __update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument
1823 struct folio *folio; local
1851 update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio, bool atomic) argument
1874 struct folio *folio, *t_folio; local
1925 struct folio *folio, *t_folio; local
1973 free_huge_folio(struct folio *folio) argument
1985 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local
1986 VM_BUG_ON_FOLIO(folio_mapcount(folio), folio); local
2017 pages_per_huge_page(h), folio); local
2019 pages_per_huge_page(h), folio); local
2050 init_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument
2059 __prep_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument
2065 prep_new_hugetlb_folio(struct hstate *h, struct folio *folio, int nid) argument
2073 __prep_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument
2150 prep_compound_gigantic_folio(struct folio *folio, unsigned int order) argument
2156 prep_compound_gigantic_folio_for_demote(struct folio *folio, unsigned int order) argument
2249 struct folio *folio; local
2283 struct folio *folio; local
2303 struct folio *folio; local
2318 struct folio *folio, *tmp_f; local
2345 struct folio *folio; local
2367 struct folio *folio = NULL; local
2404 struct folio *folio = page_folio(page); local
2515 struct folio *folio = NULL; local
2556 struct folio *folio; local
2583 struct folio *folio = NULL; local
2612 struct folio *folio; local
2634 struct folio *folio, *tmp; variable in typeref:struct:folio
2750 struct folio *folio; local
2932 restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma, unsigned long address, struct folio *folio) argument
3099 struct folio *folio = page_folio(page); local
3137 struct folio *folio; local
3258 pages_per_huge_page(h), folio); local
3335 hugetlb_folio_init_tail_vmemmap(struct folio *folio, unsigned long start_page_number, unsigned long end_page_number) argument
3355 hugetlb_folio_init_vmemmap(struct folio *folio, struct hstate *h, unsigned long nr_pages) argument
3375 struct folio *folio, *tmp_f; local
3412 struct folio *folio = (void *)page; local
3478 struct folio *folio; local
3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], local
3706 struct folio *folio, *next; local
3769 struct folio *folio; local
3936 demote_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument
4013 struct folio *folio; variable in typeref:struct:folio
6124 struct folio *folio; local
6133 hugetlb_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t idx) argument
6207 struct folio *folio; local
6439 struct folio *folio = NULL; local
6640 struct folio *folio; local
6674 struct folio *folio; local
7592 isolate_hugetlb(struct folio *folio, struct list_head *list) argument
7610 get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison) argument
7640 folio_putback_active_hugetlb(struct folio *folio) argument
[all...]
H A Dshmem.c150 struct folio **foliop, enum sgp_type sgp, gfp_t gfp,
625 struct folio *folio; local
678 folio = filemap_get_folio(inode->i_mapping, index);
679 if (IS_ERR(folio))
683 if (!folio_test_large(folio)) {
684 folio_put(folio);
695 if (!folio_trylock(folio)) {
696 folio_put(folio);
700 ret = split_folio(folio);
761 shmem_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t index, void *expected, gfp_t gfp) argument
814 shmem_delete_from_page_cache(struct folio *folio, void *radswap) argument
934 struct folio *folio; local
973 struct folio *folio; local
1083 folio); local
1278 struct folio *folio; local
1324 struct folio *folio = fbatch->folios[i]; local
1421 struct folio *folio = page_folio(page); local
1573 struct folio *folio; local
1640 struct folio *folio; local
1743 shmem_should_replace_folio(struct folio *folio, gfp_t gfp) argument
1817 shmem_set_folio_swapin_error(struct inode *inode, pgoff_t index, struct folio *folio, swp_entry_t swap) argument
1856 struct folio *folio = NULL; local
1965 struct folio *folio; local
2240 struct folio *folio = NULL; local
2618 struct folio *folio; local
2679 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local
2732 struct folio *folio; local
2764 struct folio *folio = page_folio(page); local
2800 struct folio *folio = NULL; local
2977 struct folio *folio = NULL; local
3167 struct folio *folio; local
3500 struct folio *folio; local
3564 struct folio *folio = NULL; local
4510 shmem_error_remove_folio(struct address_space *mapping, struct folio *folio) argument
4957 struct folio *folio; local
4979 struct folio *folio = shmem_read_folio_gfp(mapping, index, gfp); local
[all...]
H A Dpage_owner.c367 void __folio_copy_owner(struct folio *newfolio, struct folio *old)
395 * Do not proactively clear PAGE_EXT_OWNER{_ALLOCATED} bits as the folio
404 * We linked the original stack to the new folio, we need to do the same
405 * for the new one and the old folio otherwise there will be an imbalance
H A Dmemory-failure.c610 static void collect_procs_anon(struct folio *folio, struct page *page, argument
618 av = folio_lock_anon_vma_read(folio, NULL);
647 static void collect_procs_file(struct folio *folio, struct page *page, argument
652 struct address_space *mapping = folio->mapping;
725 static void collect_procs(struct folio *folio, struct page *page, argument
728 if (!folio->mapping)
733 collect_procs_anon(folio, pag
923 delete_from_lru_cache(struct folio *folio) argument
948 truncate_error_folio(struct folio *folio, unsigned long pfn, struct address_space *mapping) argument
1034 struct folio *folio = page_folio(p); local
1158 struct folio *folio = page_folio(p); local
1180 struct folio *folio = page_folio(p); local
1202 struct folio *folio = page_folio(p); local
1404 struct folio *folio = page_folio(page); local
1501 struct folio *folio = page_folio(page); local
1573 struct folio *folio = page_folio(hpage); local
1748 struct folio *folio = pfn_folio(pfn); local
1859 raw_hwp_list_head(struct folio *folio) argument
1868 struct folio *folio = page_folio(page); local
1899 __folio_free_raw_hwp(struct folio *folio, bool move_flag) argument
1917 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument
1960 folio_free_raw_hwp(struct folio *folio, bool move_flag) argument
1979 folio_clear_hugetlb_hwpoison(struct folio *folio) argument
2003 struct folio *folio = page_folio(page); local
2056 struct folio *folio; local
2125 folio_free_raw_hwp(struct folio *folio, bool flag) argument
2530 struct folio *folio; local
2631 mf_isolate_folio(struct folio *folio, struct list_head *pagelist) argument
2674 struct folio *folio = page_folio(page); local
[all...]
H A Dmadvise.c183 struct folio *folio; local
201 folio = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE,
203 if (folio)
204 folio_put(folio);
226 struct folio *folio; local
230 xas_for_each(&xas, folio, end_index) {
234 if (!xa_is_value(folio))
236 entry = radix_to_swp_entry(folio);
335 struct folio *folio = NULL; local
491 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local
628 struct folio *folio; local
[all...]
H A Dinternal.h56 * If a 16GB hugetlb folio were mapped by PTEs of all of its 4kB pages,
72 * the folio's entire_mapcount.
74 static inline int folio_nr_pages_mapped(struct folio *folio) argument
76 return atomic_read(&folio->_nr_pages_mapped) & FOLIO_PAGES_MAPPED;
79 static inline void *folio_raw_mapping(struct folio *folio) argument
81 unsigned long mapping = (unsigned long)folio->mapping;
107 * folio_pte_batch - detect a PTE batch for a large folio
108 * @folio
129 folio_pte_batch(struct folio *folio, unsigned long addr, pte_t *start_ptep, pte_t pte, int max_nr, fpb_t flags, bool *any_writable) argument
144 VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio); local
181 acct_reclaim_writeback(struct folio *folio) argument
250 folio_evictable(struct folio *folio) argument
276 folio_needs_release(struct folio *folio) argument
499 folio_set_order(struct folio *folio, unsigned int order) argument
514 struct folio *folio = (struct folio *)page; local
522 struct folio *folio = (struct folio *)page; local
710 folio_within_range(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
716 VM_WARN_ON_FOLIO(folio_test_ksm(folio), folio); local
738 folio_within_vma(struct folio *folio, struct vm_area_struct *vma) argument
753 mlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument
769 munlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument
872 unmap_mapping_folio(struct folio *folio) argument
873 mlock_new_folio(struct folio *folio) argument
[all...]
H A Dhuge_memory.c771 struct deferred_split *get_deferred_split_queue(struct folio *folio) argument
773 struct mem_cgroup *memcg = folio_memcg(folio);
774 struct pglist_data *pgdat = NODE_DATA(folio_nid(folio));
783 struct deferred_split *get_deferred_split_queue(struct folio *folio) argument
785 struct pglist_data *pgdat = NODE_DATA(folio_nid(folio));
791 void folio_prep_large_rmappable(struct folio *folio) argument
793 if (!folio || !folio_test_larg
800 is_transparent_hugepage(struct folio *folio) argument
872 struct folio *folio = page_folio(page); local
1000 struct folio *folio; local
1517 struct folio *folio; local
1718 struct folio *folio; local
1807 struct folio *folio; local
1910 struct folio *folio = NULL; local
2053 struct folio *folio = pfn_swap_entry_folio(entry); local
2084 struct folio *folio; local
2436 struct folio *folio; local
2649 __split_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd, unsigned long address, bool freeze, struct folio *folio) argument
2684 split_huge_pmd_address(struct vm_area_struct *vma, unsigned long address, bool freeze, struct folio *folio) argument
2730 unmap_folio(struct folio *folio) argument
2753 remap_page(struct folio *folio, unsigned long nr) argument
2792 __split_huge_page_tail(struct folio *folio, int tail, struct lruvec *lruvec, struct list_head *list, unsigned int new_order) argument
2892 struct folio *folio = page_folio(page); local
3000 can_split_folio(struct folio *folio, int *pextra_pins) argument
3042 struct folio *folio = page_folio(page); local
3230 folio_undo_large_rmappable(struct folio *folio) argument
3255 deferred_split_folio(struct folio *folio) argument
3320 struct folio *folio, *next; local
3373 struct folio *folio; local
3468 struct folio *folio; local
3547 struct folio *folio = filemap_get_folio(mapping, index); local
3676 struct folio *folio = page_folio(page); local
3725 struct folio *folio = page_folio(new); local
[all...]
H A Dgup.c53 struct folio *folio = page_folio(page); local
56 !folio_test_anon(folio))
58 if (!folio_test_large(folio) || folio_test_hugetlb(folio))
59 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page), page);
62 VM_BUG_ON_PAGE(!PageAnonExclusive(&folio->page) &&
68 * Return the folio with ref appropriately incremented,
71 static inline struct folio *try_get_folio(struct page *page, int refs)
73 struct folio *foli local
128 struct folio *folio; local
189 gup_put_folio(struct folio *folio, int refs, unsigned int flags) argument
227 struct folio *folio = page_folio(page); local
286 folio_add_pin(struct folio *folio) argument
310 struct folio *folio = page_folio(next); local
324 struct folio *folio = page_folio(list[i]); local
362 struct folio *folio; local
428 struct folio *folio; local
446 struct folio *folio; local
472 struct folio *folio; local
1307 struct folio *folio; local
2063 struct folio *folio = page_folio(pages[i]); local
2113 struct folio *folio = page_folio(pages[i]); local
2484 folio_fast_pin_allowed(struct folio *folio, unsigned int flags) argument
2593 struct folio *folio; local
2807 struct folio *folio; local
2881 struct folio *folio; local
2925 struct folio *folio; local
2971 struct folio *folio; local
[all...]
/linux-master/lib/
H A Dscatterlist.c1276 struct folio *folio; local
1285 xas_for_each(&xas, folio, ULONG_MAX) {
1286 if (xas_retry(&xas, folio))
1288 if (WARN_ON(xa_is_value(folio)))
1290 if (WARN_ON(folio_test_hugetlb(folio)))
1293 offset = offset_in_folio(folio, start);
1294 len = min_t(size_t, maxsize, folio_size(folio) - offset);
1296 sg_set_page(sg, folio_page(folio, 0), len, offset);
/linux-master/arch/xtensa/include/asm/
H A Dcacheflush.h127 void flush_dcache_folio(struct folio *folio);
/linux-master/kernel/sched/
H A Dfair.c1770 static int numa_hint_fault_latency(struct folio *folio) argument
1775 last_time = folio_xchg_access_time(folio, time);
1832 bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio, argument
1868 latency = numa_hint_fault_latency(folio);
1873 folio_nr_pages(folio));
1877 last_cpupid = folio_xchg_last_cpupid(folio, this_cpupid);
/linux-master/fs/erofs/
H A Dfscache.c57 struct folio *folio; local
65 xas_for_each(&xas, folio, last_page) {
66 if (xas_retry(&xas, folio))
69 folio_mark_uptodate(folio);
70 folio_unlock(folio);
224 static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio) argument
226 struct erofs_fscache *ctx = folio->mapping->host->i_private;
231 req = erofs_fscache_req_alloc(folio
345 erofs_fscache_read_folio(struct file *file, struct folio *folio) argument
[all...]
/linux-master/include/linux/
H A Dpage-flags.h144 * Depending on the way an anonymous folio can be mapped into a page
147 * tail pages of an anonymous folio. For now, we only expect it to be
191 /* At least one page in this folio has the hwpoison flag set */
256 * page_folio - Converts from page to folio.
259 * Every page is part of a folio. This function cannot be called on a
263 * does not hold a reference, this call may race with a folio split, so
264 * it should re-check the folio still contains this page after gaining
265 * a reference on the folio.
266 * Return: The folio which contains this page.
269 const struct page *: (const struct folio *)_compound_hea
308 const_folio_flags(const struct folio *folio, unsigned n) argument
318 folio_flags(struct folio *folio, unsigned n) argument
581 folio_test_swapcache(const struct folio *folio) argument
686 folio_mapping_flags(const struct folio *folio) argument
696 folio_test_anon(const struct folio *folio) argument
706 __folio_test_movable(const struct folio *folio) argument
725 folio_test_ksm(const struct folio *folio) argument
754 folio_xor_flags_has_waiters(struct folio *folio, unsigned long mask) argument
770 folio_test_uptodate(const struct folio *folio) argument
792 __folio_mark_uptodate(struct folio *folio) argument
798 folio_mark_uptodate(struct folio *folio) argument
829 folio_test_head(const struct folio *folio) argument
850 folio_test_large(const struct folio *folio) argument
1157 folio_has_private(const struct folio *folio) argument
[all...]
H A Dmm.h215 #define folio_page_idx(folio, p) (page_to_pfn(p) - folio_pfn(folio))
218 #define folio_page_idx(folio, p) ((p) - &(folio)->page)
230 static inline struct folio *lru_to_folio(struct list_head *head)
232 return list_entry((head)->prev, struct folio, lru);
1091 struct folio *folio = (struct folio *)page; local
1093 if (!test_bit(PG_head, &folio
1107 folio_order(struct folio *folio) argument
1138 folio_put_testzero(struct folio *folio) argument
1202 folio_entire_mapcount(struct folio *folio) argument
1255 folio_mapcount(struct folio *folio) argument
1262 folio_large_is_mapped(struct folio *folio) argument
1278 folio_mapped(struct folio *folio) argument
1472 folio_get(struct folio *folio) argument
1474 VM_BUG_ON_FOLIO(folio_ref_zero_or_close_to_overflow(folio), folio); local
1505 folio_put(struct folio *folio) argument
1525 folio_put_refs(struct folio *folio, int refs) argument
1572 struct folio *folio = page_folio(page); local
1668 folio_nid(const struct folio *folio) argument
1723 folio_xchg_last_cpupid(struct folio *folio, int cpupid) argument
1728 folio_last_cpupid(struct folio *folio) argument
1737 folio_last_cpupid(struct folio *folio) argument
1750 folio_xchg_access_time(struct folio *folio, int time) argument
1769 folio_xchg_last_cpupid(struct folio *folio, int cpupid) argument
1774 folio_xchg_access_time(struct folio *folio, int time) argument
1779 folio_last_cpupid(struct folio *folio) argument
1887 folio_zone(const struct folio *folio) argument
1892 folio_pgdat(const struct folio *folio) argument
1919 folio_pfn(struct folio *folio) argument
1954 folio_maybe_dma_pinned(struct folio *folio) argument
1982 folio_needs_cow_for_dma(struct vm_area_struct *vma, struct folio *folio) argument
2010 is_zero_folio(const struct folio *folio) argument
2017 folio_is_longterm_pinnable(struct folio *folio) argument
2038 folio_is_longterm_pinnable(struct folio *folio) argument
2072 folio_nr_pages(struct folio *folio) argument
2097 struct folio *folio = (struct folio *)page; local
2131 folio_next(struct folio *folio) argument
2148 folio_shift(struct folio *folio) argument
2161 folio_size(struct folio *folio) argument
2179 folio_estimated_sharers(struct folio *folio) argument
2192 arch_make_folio_accessible(struct folio *folio) argument
2245 folio_address(const struct folio *folio) argument
2283 folio_is_pfmemalloc(const struct folio *folio) argument
2621 mm_counter_file(struct folio *folio) argument
2628 mm_counter(struct folio *folio) argument
2949 struct folio *folio = ptdesc_folio(ptdesc); local
2960 struct folio *folio = ptdesc_folio(ptdesc); local
3064 struct folio *folio = ptdesc_folio(ptdesc); local
3075 struct folio *folio = ptdesc_folio(ptdesc); local
3103 struct folio *folio = ptdesc_folio(ptdesc); local
3111 struct folio *folio = ptdesc_folio(ptdesc); local
[all...]
H A Dswapops.h500 static inline struct folio *pfn_swap_entry_folio(swp_entry_t entry)
502 struct folio *folio = pfn_folio(swp_offset_pfn(entry)); local
506 * corresponding folio is locked
508 BUG_ON(is_migration_entry(entry) && !folio_test_locked(folio));
510 return folio;
H A Dshmem_fs.h144 int shmem_get_folio(struct inode *inode, pgoff_t index, struct folio **foliop,
146 struct folio *shmem_read_folio_gfp(struct address_space *mapping,
149 static inline struct folio *shmem_read_folio(struct address_space *mapping,
193 struct folio **foliop);
/linux-master/fs/ntfs3/
H A Dinode.c565 struct folio *folio = bh->b_folio; local
580 err = attr_data_read_resident(ni, &folio->page);
653 if (folio) {
658 folio_set_bh(bh, folio, off);
664 folio_zero_segment(folio, off + voff, off + block_size);
705 static int ntfs_read_folio(struct file *file, struct folio *folio) argument
707 struct page *page = &folio->page;
731 return mpage_read_folio(folio, ntfs_get_bloc
857 ntfs_resident_writepage(struct folio *folio, struct writeback_control *wbc, void *data) argument
[all...]
H A Dfile.c190 struct folio *folio; local
199 folio = __filemap_get_folio(mapping, idx,
202 if (IS_ERR(folio))
203 return PTR_ERR(folio);
205 head = folio_buffers(folio);
207 head = create_empty_buffers(folio, blocksize, 0);
225 if (folio_test_uptodate(folio))
229 folio_unlock(folio);
230 folio_put(folio);
[all...]
/linux-master/fs/netfs/
H A Dbuffered_write.c21 NETFS_JUST_PREFETCH, /* We have to read the folio anyway */
22 NETFS_WHOLE_FOLIO_MODIFY, /* We're going to overwrite the whole folio */
31 static void netfs_set_group(struct folio *folio, struct netfs_group *netfs_group) argument
33 if (netfs_group && !folio_get_private(folio))
34 folio_attach_private(folio, netfs_get_group(netfs_group));
38 static void netfs_folio_start_fscache(bool caching, struct folio *folio) argument
41 folio_start_fscache(folio);
44 static void netfs_folio_start_fscache(bool caching, struct folio *foli argument
56 netfs_how_to_modify(struct netfs_inode *ctx, struct file *file, struct folio *folio, void *netfs_group, size_t flen, size_t offset, size_t len, bool maybe_trouble) argument
163 struct folio *folio; local
508 struct folio *folio = page_folio(vmf->page); local
567 struct folio *folio; local
605 struct folio *folio; local
642 struct folio *folio; local
781 struct folio *folio; local
892 netfs_write_back_from_locked_folio(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, struct folio *folio, unsigned long long start, unsigned long long end) argument
1006 struct folio *folio; local
1204 netfs_launder_folio(struct folio *folio) argument
[all...]
/linux-master/arch/arc/mm/
H A Dtlb.c490 struct folio *folio = page_folio(page); local
491 int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags);
493 unsigned long offset = offset_in_folio(folio, paddr);
494 nr = folio_nr_pages(folio);
/linux-master/fs/btrfs/
H A Dinode.c4740 struct folio *folio; local
4772 folio = __filemap_get_folio(mapping, index,
4774 if (IS_ERR(folio)) {
4782 if (!folio_test_uptodate(folio)) {
4783 ret = btrfs_read_folio(NULL, folio);
4784 folio_lock(folio);
4785 if (folio->mapping != mapping) {
4786 folio_unlock(folio);
4787 folio_put(folio);
7947 struct folio *folio = page_folio(page); local
7971 __btrfs_release_folio(struct folio *folio, gfp_t gfp_flags) argument
7982 btrfs_release_folio(struct folio *folio, gfp_t gfp_flags) argument
8010 btrfs_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument
8195 struct folio *folio = page_folio(page); local
[all...]
/linux-master/fs/smb/client/
H A Dsmb2ops.c4311 struct folio *folio; local
4316 xas_for_each_marked(&xas, folio, ULONG_MAX, XA_MARK_0) {
4317 folio_put(folio);
H A Dcifsfs.c1237 * Flush out either the folio that overlaps the beginning of a range in which
1238 * pos resides or the folio that overlaps the end of a range unless that folio
1240 * bounds to encompass the folio.
1245 struct folio *folio; local
1251 folio = filemap_get_folio(inode->i_mapping, index);
1252 if (IS_ERR(folio))
1255 size = folio_size(folio);
1256 fpos = folio_pos(folio);
[all...]

Completed in 523 milliseconds

1234567891011>>