/linux-master/mm/ |
H A D | hugetlb_vmemmap.h | 21 int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio); 25 void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio); 46 static inline int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio) argument 59 static inline void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio) argument
|
H A D | debug.c | 54 static void __dump_folio(struct folio *folio, struct page *page, argument 57 struct address_space *mapping = folio_mapping(folio); 65 if (!folio_test_slab(folio)) { 67 if (folio_test_large(folio)) 68 mapcount += folio_entire_mapcount(folio); 72 folio_ref_count(folio), mapcount, mapping, 73 folio->index + idx, pfn); 74 if (folio_test_large(folio)) { 76 folio_order(folio), 115 struct folio *foliop, folio; local [all...] |
H A D | internal.h | 56 * If a 16GB hugetlb folio were mapped by PTEs of all of its 4kB pages, 72 * the folio's entire_mapcount. 74 static inline int folio_nr_pages_mapped(struct folio *folio) argument 76 return atomic_read(&folio->_nr_pages_mapped) & FOLIO_PAGES_MAPPED; 79 static inline void *folio_raw_mapping(struct folio *folio) argument 81 unsigned long mapping = (unsigned long)folio->mapping; 107 * folio_pte_batch - detect a PTE batch for a large folio 108 * @folio 129 folio_pte_batch(struct folio *folio, unsigned long addr, pte_t *start_ptep, pte_t pte, int max_nr, fpb_t flags, bool *any_writable) argument 144 VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio); local 181 acct_reclaim_writeback(struct folio *folio) argument 250 folio_evictable(struct folio *folio) argument 276 folio_needs_release(struct folio *folio) argument 499 folio_set_order(struct folio *folio, unsigned int order) argument 514 struct folio *folio = (struct folio *)page; local 522 struct folio *folio = (struct folio *)page; local 710 folio_within_range(struct folio *folio, struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 716 VM_WARN_ON_FOLIO(folio_test_ksm(folio), folio); local 738 folio_within_vma(struct folio *folio, struct vm_area_struct *vma) argument 753 mlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 769 munlock_vma_folio(struct folio *folio, struct vm_area_struct *vma) argument 872 unmap_mapping_folio(struct folio *folio) argument 873 mlock_new_folio(struct folio *folio) argument [all...] |
H A D | workingset.c | 232 static void *lru_gen_eviction(struct folio *folio) argument 239 int type = folio_is_file_lru(folio); 240 int delta = folio_nr_pages(folio); 241 int refs = folio_lru_refs(folio); 243 struct mem_cgroup *memcg = folio_memcg(folio); 244 struct pglist_data *pgdat = folio_pgdat(folio); 260 * Tests if the shadow entry is for a folio that was recently evicted. 280 static void lru_gen_refault(struct folio *folio, voi argument 329 lru_gen_eviction(struct folio *folio) argument 340 lru_gen_refault(struct folio *folio, void *shadow) argument 382 workingset_eviction(struct folio *folio, struct mem_cgroup *target_memcg) argument 390 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 391 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local 530 workingset_refault(struct folio *folio, void *shadow) argument 584 workingset_activation(struct folio *folio) argument [all...] |
H A D | memory-failure.c | 610 static void collect_procs_anon(struct folio *folio, struct page *page, argument 618 av = folio_lock_anon_vma_read(folio, NULL); 647 static void collect_procs_file(struct folio *folio, struct page *page, argument 652 struct address_space *mapping = folio->mapping; 725 static void collect_procs(struct folio *folio, struct page *page, argument 728 if (!folio->mapping) 733 collect_procs_anon(folio, pag 923 delete_from_lru_cache(struct folio *folio) argument 948 truncate_error_folio(struct folio *folio, unsigned long pfn, struct address_space *mapping) argument 1034 struct folio *folio = page_folio(p); local 1158 struct folio *folio = page_folio(p); local 1180 struct folio *folio = page_folio(p); local 1202 struct folio *folio = page_folio(p); local 1404 struct folio *folio = page_folio(page); local 1501 struct folio *folio = page_folio(page); local 1573 struct folio *folio = page_folio(hpage); local 1748 struct folio *folio = pfn_folio(pfn); local 1859 raw_hwp_list_head(struct folio *folio) argument 1868 struct folio *folio = page_folio(page); local 1899 __folio_free_raw_hwp(struct folio *folio, bool move_flag) argument 1917 folio_set_hugetlb_hwpoison(struct folio *folio, struct page *page) argument 1960 folio_free_raw_hwp(struct folio *folio, bool move_flag) argument 1979 folio_clear_hugetlb_hwpoison(struct folio *folio) argument 2003 struct folio *folio = page_folio(page); local 2056 struct folio *folio; local 2125 folio_free_raw_hwp(struct folio *folio, bool flag) argument 2530 struct folio *folio; local 2631 mf_isolate_folio(struct folio *folio, struct list_head *pagelist) argument 2674 struct folio *folio = page_folio(page); local [all...] |
H A D | hugetlb.c | 59 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument 61 return cma_pages_valid(hugetlb_cma[folio_nid(folio)], &folio->page, 65 static bool hugetlb_cma_folio(struct folio *folio, unsigned int order) argument 1311 static void enqueue_hugetlb_folio(struct hstate *h, struct folio *folio) argument 1313 int nid = folio_nid(folio); 1316 VM_BUG_ON_FOLIO(folio_ref_count(folio), foli local 1327 struct folio *folio; local 1363 struct folio *folio; local 1395 struct folio *folio = NULL; local 1512 __destroy_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument 1535 destroy_compound_hugetlb_folio_for_demote(struct folio *folio, unsigned int order) argument 1542 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 1548 free_gigantic_folio(struct folio *folio, unsigned int order) argument 1616 free_gigantic_folio(struct folio *folio, unsigned int order) argument 1618 destroy_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 1622 __clear_hugetlb_destructor(struct hstate *h, struct folio *folio) argument 1640 __remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus, bool demote) argument 1646 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio(folio), folio); local 1647 VM_BUG_ON_FOLIO(hugetlb_cgroup_from_folio_rsvd(folio), folio); local 1683 remove_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1689 remove_hugetlb_folio_for_demote(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1695 add_hugetlb_folio(struct hstate *h, struct folio *folio, bool adjust_surplus) argument 1741 __update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument 1823 struct folio *folio; local 1851 update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio, bool atomic) argument 1874 struct folio *folio, *t_folio; local 1925 struct folio *folio, *t_folio; local 1973 free_huge_folio(struct folio *folio) argument 1985 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); local 1986 VM_BUG_ON_FOLIO(folio_mapcount(folio), folio); local 2017 pages_per_huge_page(h), folio); local 2019 pages_per_huge_page(h), folio); local 2050 init_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument 2059 __prep_new_hugetlb_folio(struct hstate *h, struct folio *folio) argument 2065 prep_new_hugetlb_folio(struct hstate *h, struct folio *folio, int nid) argument 2073 __prep_compound_gigantic_folio(struct folio *folio, unsigned int order, bool demote) argument 2150 prep_compound_gigantic_folio(struct folio *folio, unsigned int order) argument 2156 prep_compound_gigantic_folio_for_demote(struct folio *folio, unsigned int order) argument 2249 struct folio *folio; local 2283 struct folio *folio; local 2303 struct folio *folio; local 2318 struct folio *folio, *tmp_f; local 2345 struct folio *folio; local 2367 struct folio *folio = NULL; local 2404 struct folio *folio = page_folio(page); local 2515 struct folio *folio = NULL; local 2556 struct folio *folio; local 2583 struct folio *folio = NULL; local 2612 struct folio *folio; local 2634 struct folio *folio, *tmp; variable in typeref:struct:folio 2750 struct folio *folio; local 2932 restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma, unsigned long address, struct folio *folio) argument 3099 struct folio *folio = page_folio(page); local 3137 struct folio *folio; local 3258 pages_per_huge_page(h), folio); local 3335 hugetlb_folio_init_tail_vmemmap(struct folio *folio, unsigned long start_page_number, unsigned long end_page_number) argument 3355 hugetlb_folio_init_vmemmap(struct folio *folio, struct hstate *h, unsigned long nr_pages) argument 3375 struct folio *folio, *tmp_f; local 3412 struct folio *folio = (void *)page; local 3478 struct folio *folio; local 3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], local 3706 struct folio *folio, *next; local 3769 struct folio *folio; local 3936 demote_free_hugetlb_folio(struct hstate *h, struct folio *folio) argument 4013 struct folio *folio; variable in typeref:struct:folio 6124 struct folio *folio; local 6133 hugetlb_add_to_page_cache(struct folio *folio, struct address_space *mapping, pgoff_t idx) argument 6207 struct folio *folio; local 6439 struct folio *folio = NULL; local 6640 struct folio *folio; local 6674 struct folio *folio; local 7592 isolate_hugetlb(struct folio *folio, struct list_head *list) argument 7610 get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison) argument 7640 folio_putback_active_hugetlb(struct folio *folio) argument [all...] |
/linux-master/arch/openrisc/mm/ |
H A D | cache.c | 46 struct folio *folio = page_folio(pfn_to_page(pfn)); local 47 int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags); 55 unsigned int nr = folio_nr_pages(folio); 58 sync_icache_dcache(folio_page(folio, nr));
|
/linux-master/include/linux/ |
H A D | highmem.h | 99 * kmap_local_folio - Map a page in this folio for temporary usage 100 * @folio: The folio containing the page. 101 * @offset: The byte offset within the folio which identifies the page. 132 static inline void *kmap_local_folio(struct folio *folio, size_t offset); 220 * Return: A folio containing one allocated and zeroed page or NULL if 224 struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma, 227 struct folio *folio; local 449 memcpy_from_folio(char *to, struct folio *folio, size_t offset, size_t len) argument 477 memcpy_to_folio(struct folio *folio, size_t offset, const char *from, size_t len) argument 514 folio_zero_tail(struct folio *folio, size_t offset, void *kaddr) argument 550 folio_fill_tail(struct folio *folio, size_t offset, const char *from, size_t len) argument 588 memcpy_from_file_folio(char *to, struct folio *folio, loff_t pos, size_t len) argument 614 folio_zero_segments(struct folio *folio, size_t start1, size_t xend1, size_t start2, size_t xend2) argument 626 folio_zero_segment(struct folio *folio, size_t start, size_t xend) argument 638 folio_zero_range(struct folio *folio, size_t start, size_t length) argument 653 folio_release_kmap(struct folio *folio, void *addr) argument [all...] |
H A D | memcontrol.h | 362 static inline bool folio_memcg_kmem(struct folio *folio); 377 * __folio_memcg - Get the memory cgroup associated with a non-kmem folio 378 * @folio: Pointer to the folio. 380 * Returns a pointer to the memory cgroup associated with the folio, 381 * or NULL. This function assumes that the folio is known to have a 386 static inline struct mem_cgroup *__folio_memcg(struct folio *folio) argument 388 unsigned long memcg_data = folio 390 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local 407 __folio_objcg(struct folio *folio) argument 411 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local 439 folio_memcg(struct folio *folio) argument 462 folio_memcg_rcu(struct folio *folio) argument 466 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); local 501 folio_memcg_check(struct folio *folio) argument 552 folio_memcg_kmem(struct folio *folio) argument 561 folio_memcg_kmem(struct folio *folio) argument 686 mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp) argument 709 mem_cgroup_uncharge(struct folio *folio) argument 770 folio_lruvec(struct folio *folio) argument 793 lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) argument 1100 count_memcg_folio_events(struct folio *folio, enum vm_event_item idx, unsigned long nr) argument 1174 folio_memcg(struct folio *folio) argument 1184 folio_memcg_rcu(struct folio *folio) argument 1190 folio_memcg_check(struct folio *folio) argument 1205 folio_memcg_kmem(struct folio *folio) argument 1265 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 1270 mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp) argument 1282 mem_cgroup_swapin_charge_folio(struct folio *folio, struct mm_struct *mm, gfp_t gfp, swp_entry_t entry) argument 1292 mem_cgroup_uncharge(struct folio *folio) argument 1320 folio_lruvec(struct folio *folio) argument 1327 lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) argument 1376 folio_lruvec_lock(struct folio *folio) argument 1384 folio_lruvec_lock_irq(struct folio *folio) argument 1392 folio_lruvec_lock_irqsave(struct folio *folio, unsigned long *flagsp) argument 1485 folio_memcg_lock(struct folio *folio) argument 1489 folio_memcg_unlock(struct folio *folio) argument 1612 count_memcg_folio_events(struct folio *folio, enum vm_event_item idx, unsigned long nr) argument 1675 folio_matches_lruvec(struct folio *folio, struct lruvec *lruvec) argument 1683 folio_lruvec_relock_irq(struct folio *folio, struct lruvec *locked_lruvec) argument 1697 folio_lruvec_relock_irqsave(struct folio *folio, struct lruvec **lruvecp, unsigned long *flags) argument 1720 mem_cgroup_track_foreign_dirty(struct folio *folio, struct bdi_writeback *wb) argument 1750 mem_cgroup_track_foreign_dirty(struct folio *folio, struct bdi_writeback *wb) argument 1908 get_obj_cgroup_from_folio(struct folio *folio) argument [all...] |
H A D | vmstat.h | 418 static inline void __zone_stat_mod_folio(struct folio *folio, argument 421 __mod_zone_page_state(folio_zone(folio), item, nr); 424 static inline void __zone_stat_add_folio(struct folio *folio, argument 427 __mod_zone_page_state(folio_zone(folio), item, folio_nr_pages(folio)); 430 static inline void __zone_stat_sub_folio(struct folio *folio, argument 433 __mod_zone_page_state(folio_zone(folio), ite 436 zone_stat_mod_folio(struct folio *folio, enum zone_stat_item item, long nr) argument 442 zone_stat_add_folio(struct folio *folio, enum zone_stat_item item) argument 448 zone_stat_sub_folio(struct folio *folio, enum zone_stat_item item) argument 454 __node_stat_mod_folio(struct folio *folio, enum node_stat_item item, long nr) argument 460 __node_stat_add_folio(struct folio *folio, enum node_stat_item item) argument 466 __node_stat_sub_folio(struct folio *folio, enum node_stat_item item) argument 472 node_stat_mod_folio(struct folio *folio, enum node_stat_item item, long nr) argument 478 node_stat_add_folio(struct folio *folio, enum node_stat_item item) argument 484 node_stat_sub_folio(struct folio *folio, enum node_stat_item item) argument 562 lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument 592 __lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument 598 lruvec_stat_mod_folio(struct folio *folio, enum node_stat_item idx, int val) argument 612 __lruvec_stat_add_folio(struct folio *folio, enum node_stat_item idx) argument 618 __lruvec_stat_sub_folio(struct folio *folio, enum node_stat_item idx) argument 624 lruvec_stat_add_folio(struct folio *folio, enum node_stat_item idx) argument 630 lruvec_stat_sub_folio(struct folio *folio, enum node_stat_item idx) argument [all...] |
H A D | netfs.h | 28 #define folio_test_fscache(folio) folio_test_private_2(folio) 36 * folio_start_fscache - Start an fscache write on a folio. 37 * @folio: The folio. 39 * Call this function before writing a folio to a local cache. Starting a 42 static inline void folio_start_fscache(struct folio *folio) argument 44 VM_BUG_ON_FOLIO(folio_test_private_2(folio), folio); local 56 folio_end_fscache(struct folio *folio) argument 69 folio_wait_fscache(struct folio *folio) argument 87 folio_wait_fscache_killable(struct folio *folio) argument 169 netfs_folio_info(struct folio *folio) argument 178 netfs_folio_group(struct folio *folio) argument [all...] |
H A D | page-flags.h | 144 * Depending on the way an anonymous folio can be mapped into a page 147 * tail pages of an anonymous folio. For now, we only expect it to be 191 /* At least one page in this folio has the hwpoison flag set */ 256 * page_folio - Converts from page to folio. 259 * Every page is part of a folio. This function cannot be called on a 263 * does not hold a reference, this call may race with a folio split, so 264 * it should re-check the folio still contains this page after gaining 265 * a reference on the folio. 266 * Return: The folio which contains this page. 269 const struct page *: (const struct folio *)_compound_hea 308 const_folio_flags(const struct folio *folio, unsigned n) argument 318 folio_flags(struct folio *folio, unsigned n) argument 581 folio_test_swapcache(const struct folio *folio) argument 686 folio_mapping_flags(const struct folio *folio) argument 696 folio_test_anon(const struct folio *folio) argument 706 __folio_test_movable(const struct folio *folio) argument 725 folio_test_ksm(const struct folio *folio) argument 754 folio_xor_flags_has_waiters(struct folio *folio, unsigned long mask) argument 770 folio_test_uptodate(const struct folio *folio) argument 792 __folio_mark_uptodate(struct folio *folio) argument 798 folio_mark_uptodate(struct folio *folio) argument 829 folio_test_head(const struct folio *folio) argument 850 folio_test_large(const struct folio *folio) argument 1157 folio_has_private(const struct folio *folio) argument [all...] |
/linux-master/fs/affs/ |
H A D | symlink.c | 14 static int affs_symlink_read_folio(struct file *file, struct folio *folio) argument 17 struct inode *inode = folio->mapping->host; 18 char *link = folio_address(folio); 60 folio_mark_uptodate(folio); 61 folio_unlock(folio); 64 folio_unlock(folio);
|
/linux-master/fs/nilfs2/ |
H A D | file.c | 48 struct folio *folio = page_folio(vmf->page); local 58 folio_lock(folio); 59 if (folio->mapping != inode->i_mapping || 60 folio_pos(folio) >= i_size_read(inode) || 61 !folio_test_uptodate(folio)) { 62 folio_unlock(folio); 68 * check to see if the folio is mapped already (no holes) 70 if (folio_test_mappedtodisk(folio)) 73 head = folio_buffers(folio); [all...] |
/linux-master/fs/xfs/scrub/ |
H A D | xfile.h | 24 #define XFILE_ALLOC (1 << 0) /* allocate folio if not present */ 25 struct folio *xfile_get_folio(struct xfile *xf, loff_t offset, size_t len, 27 void xfile_put_folio(struct xfile *xf, struct folio *folio);
|
/linux-master/arch/nios2/mm/ |
H A D | cacheflush.c | 74 static void flush_aliases(struct address_space *mapping, struct folio *folio) argument 80 unsigned long nr = folio_nr_pages(folio); 82 pgoff = folio->index; 163 static void __flush_dcache_folio(struct folio *folio) argument 170 unsigned long start = (unsigned long)folio_address(folio); 172 __flush_dcache(start, start + folio_size(folio)); 175 void flush_dcache_folio(struct folio *folio) argument 214 struct folio *folio; local [all...] |
/linux-master/fs/bcachefs/ |
H A D | fs-io-pagecache.c | 19 struct folio *f; 126 struct bch_folio *__bch2_folio_create(struct folio *folio, gfp_t gfp) argument 132 folio_sectors(folio), gfp); 137 folio_attach_private(folio, s); 141 struct bch_folio *bch2_folio_create(struct folio *folio, gfp_t gfp) argument 143 return bch2_folio(folio) ?: __bch2_folio_create(folio, gfp); 155 static void __bch2_folio_set(struct folio *foli argument 222 struct folio *folio = fs[folio_idx]; local 286 struct folio *folio = fbatch.folios[i]; local 331 struct folio *folio = fbatch.folios[i]; local 380 bch2_get_folio_disk_reservation(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, bool check_enospc) argument 422 bch2_folio_reservation_get(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, struct bch2_folio_reservation *res, unsigned offset, unsigned len) argument 468 bch2_clear_folio_bits(struct folio *folio) argument 497 bch2_set_folio_dirty(struct bch_fs *c, struct bch_inode_info *inode, struct folio *folio, struct bch2_folio_reservation *res, unsigned offset, unsigned len) argument 581 struct folio *folio = page_folio(vmf->page); local 634 bch2_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 642 bch2_release_folio(struct folio *folio, gfp_t gfp_mask) argument 653 folio_data_offset(struct folio *folio, loff_t pos, unsigned min_replicas) argument 687 struct folio *folio = fbatch.folios[i]; local 726 struct folio *folio; local [all...] |
H A D | fs-io-buffered.h | 7 int bch2_read_single_folio(struct folio *, struct address_space *); 8 int bch2_read_folio(struct file *, struct folio *);
|
/linux-master/fs/ |
H A D | mpage.c | 53 folio_set_error(fi.folio); 55 folio_mark_uptodate(fi.folio); 56 folio_unlock(fi.folio); 69 folio_set_error(fi.folio); 70 mapping_set_error(fi.folio->mapping, err); 72 folio_end_writeback(fi.folio); 104 static void map_buffer_to_folio(struct folio *folio, struct buffer_head *bh, argument 107 struct inode *inode = folio->mapping->host; 111 head = folio_buffers(folio); 140 struct folio *folio; member in struct:mpage_readpage_args 160 struct folio *folio = args->folio; local 181 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); local 371 struct folio *folio; local 391 mpage_read_folio(struct folio *folio, get_block_t get_block) argument 433 clean_buffers(struct folio *folio, unsigned first_unmapped) argument 458 __mpage_writepage(struct folio *folio, struct writeback_control *wbc, void *data) argument [all...] |
/linux-master/include/trace/events/ |
H A D | filemap.h | 18 TP_PROTO(struct folio *folio), 20 TP_ARGS(folio), 31 __entry->pfn = folio_pfn(folio); 32 __entry->i_ino = folio->mapping->host->i_ino; 33 __entry->index = folio->index; 34 if (folio->mapping->host->i_sb) 35 __entry->s_dev = folio->mapping->host->i_sb->s_dev; 37 __entry->s_dev = folio->mapping->host->i_rdev; 38 __entry->order = folio_order(folio); [all...] |
/linux-master/include/linux/sched/ |
H A D | numa_balancing.h | 33 bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio, 51 struct folio *folio, int src_nid, int dst_cpu) 50 should_numa_migrate_memory(struct task_struct *p, struct folio *folio, int src_nid, int dst_cpu) argument
|
/linux-master/arch/riscv/include/asm/ |
H A D | cacheflush.h | 18 static inline void flush_dcache_folio(struct folio *folio) argument 20 if (test_bit(PG_dcache_clean, &folio->flags)) 21 clear_bit(PG_dcache_clean, &folio->flags);
|
/linux-master/arch/mips/include/asm/ |
H A D | cacheflush.h | 39 #define folio_test_dcache_dirty(folio) \ 40 test_bit(PG_dcache_dirty, &(folio)->flags) 41 #define folio_set_dcache_dirty(folio) \ 42 set_bit(PG_dcache_dirty, &(folio)->flags) 43 #define folio_clear_dcache_dirty(folio) \ 44 clear_bit(PG_dcache_dirty, &(folio)->flags) 56 static inline void flush_dcache_folio(struct folio *folio) argument 59 __flush_dcache_pages(&folio->page, folio_nr_pages(folio)); [all...] |
/linux-master/fs/afs/ |
H A D | dir_edit.c | 108 * Get a new directory folio. 110 static struct folio *afs_dir_get_folio(struct afs_vnode *vnode, pgoff_t index) 113 struct folio *folio; local 115 folio = __filemap_get_folio(mapping, index, 118 if (IS_ERR(folio)) { 122 if (!folio_test_private(folio)) 123 folio_attach_private(folio, (void *)1); 124 return folio; 212 struct folio *folio local 379 struct folio *folio0, *folio; local [all...] |
/linux-master/fs/nfs/ |
H A D | fscache.h | 87 extern int nfs_netfs_folio_unlock(struct folio *folio); 100 extern int nfs_netfs_read_folio(struct file *file, struct folio *folio); 102 static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument 104 if (folio_test_fscache(folio)) { 107 folio_wait_fscache(folio); 109 fscache_note_page_release(netfs_i_cookie(netfs_inode(folio->mapping->host))); 166 static inline int nfs_netfs_folio_unlock(struct folio *foli argument 181 nfs_netfs_read_folio(struct file *file, struct folio *folio) argument 186 nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument [all...] |