/linux-master/tools/virtio/linux/ |
H A D | mm_types.h | 1 struct folio { struct
|
/linux-master/mm/ |
H A D | swap.c | 60 * The following folio batches are grouped together because they are protected 77 static void __page_cache_release(struct folio *folio, struct lruvec **lruvecp, argument 80 if (folio_test_lru(folio)) { 81 folio_lruvec_relock_irqsave(folio, lruvecp, flagsp); 82 lruvec_del_folio(*lruvecp, folio); 83 __folio_clear_lru_flags(folio); 92 if (unlikely(folio_test_mlocked(folio))) { 93 long nr_pages = folio_nr_pages(folio); 95 __folio_clear_mlocked(folio); 105 page_cache_release(struct folio *folio) argument 115 __folio_put(struct folio *folio) argument 142 struct folio *folio, *next; local 166 lru_add_fn(struct lruvec *lruvec, struct folio *folio) argument 171 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 213 struct folio *folio = fbatch->folios[i]; local 230 folio_batch_add_and_move(struct folio_batch *fbatch, struct folio *folio, move_fn_t move_fn) argument 239 lru_move_tail_fn(struct lruvec *lruvec, struct folio *folio) argument 256 folio_rotate_reclaimable(struct folio *folio) argument 323 lru_note_cost_refault(struct folio *folio) argument 329 folio_activate_fn(struct lruvec *lruvec, struct folio *folio) argument 354 folio_activate(struct folio *folio) argument 373 folio_activate(struct folio *folio) argument 386 __lru_cache_activate_folio(struct folio *folio) argument 417 folio_inc_refs(struct folio *folio) argument 445 folio_inc_refs(struct folio *folio) argument 463 folio_mark_accessed(struct folio *folio) argument 506 folio_add_lru(struct folio *folio) argument 512 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 535 folio_add_lru_vma(struct folio *folio, struct vm_area_struct *vma) argument 537 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); local 566 lru_deactivate_file_fn(struct lruvec *lruvec, struct folio *folio) argument 607 lru_deactivate_fn(struct lruvec *lruvec, struct folio *folio) argument 623 lru_lazyfree_fn(struct lruvec *lruvec, struct folio *folio) argument 695 deactivate_file_folio(struct folio *folio) argument 718 folio_deactivate(struct folio *folio) argument 739 folio_mark_lazyfree(struct folio *folio) argument 975 struct folio *folio = folios->folios[i]; local 1050 struct folio *folio = page_folio(encoded_page_ptr(encoded[i])); local 1102 struct folio *folio = fbatch->folios[i]; local [all...] |
H A D | page_idle.c | 34 static struct folio *page_idle_get_folio(unsigned long pfn) 37 struct folio *folio; local 42 folio = page_folio(page); 43 if (!folio_test_lru(folio) || !folio_try_get(folio)) 45 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) { 46 folio_put(folio); 47 folio 52 page_idle_clear_pte_refs_one(struct folio *folio, struct vm_area_struct *vma, unsigned long addr, void *arg) argument 89 page_idle_clear_pte_refs(struct folio *folio) argument 119 struct folio *folio; local 164 struct folio *folio; local [all...] |
H A D | swap.h | 13 void swap_read_folio(struct folio *folio, bool do_poll, 23 void __swap_writepage(struct folio *folio, struct writeback_control *wbc); 35 bool add_to_swap(struct folio *folio); 37 int add_to_swap_cache(struct folio *folio, swp_entry_t entry, 39 void __delete_from_swap_cache(struct folio *folio, 61 folio_swap_flags(struct folio *folio) argument 67 swap_read_folio(struct folio *folio, bool do_poll, struct swap_iocb **plug) argument 118 add_to_swap(struct folio *folio) argument 128 add_to_swap_cache(struct folio *folio, swp_entry_t entry, gfp_t gfp_mask, void **shadowp) argument 134 __delete_from_swap_cache(struct folio *folio, swp_entry_t entry, void *shadow) argument 139 delete_from_swap_cache(struct folio *folio) argument 148 folio_swap_flags(struct folio *folio) argument [all...] |
H A D | truncate.c | 83 struct folio *folio = fbatch->folios[i]; local 86 if (!xa_is_value(folio)) { 87 fbatch->folios[j++] = folio; 96 __clear_shadow_entry(mapping, index, folio); 139 * folio_invalidate - Invalidate part or all of a folio. 140 * @folio: The folio which is affected. 144 * folio_invalidate() is called when all or part of the folio has become 153 void folio_invalidate(struct folio *foli argument 172 truncate_cleanup_folio(struct folio *folio) argument 189 truncate_inode_folio(struct address_space *mapping, struct folio *folio) argument 210 truncate_inode_partial_folio(struct folio *folio, loff_t start, loff_t end) argument 253 generic_error_remove_folio(struct address_space *mapping, struct folio *folio) argument 279 mapping_evict_folio(struct address_space *mapping, struct folio *folio) argument 329 struct folio *folio; local 405 struct folio *folio = fbatch.folios[i]; local 501 struct folio *folio = fbatch.folios[i]; local 560 invalidate_complete_folio2(struct address_space *mapping, struct folio *folio) argument 589 folio_launder(struct address_space *mapping, struct folio *folio) argument 627 struct folio *folio = fbatch.folios[i]; local 784 struct folio *folio; local [all...] |
/linux-master/include/trace/events/ |
H A D | pagemap.h | 19 #define trace_pagemap_flags(folio) ( \ 20 (folio_test_anon(folio) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \ 21 (folio_mapped(folio) ? PAGEMAP_MAPPED : 0) | \ 22 (folio_test_swapcache(folio) ? PAGEMAP_SWAPCACHE : 0) | \ 23 (folio_test_swapbacked(folio) ? PAGEMAP_SWAPBACKED : 0) | \ 24 (folio_test_mappedtodisk(folio) ? PAGEMAP_MAPPEDDISK : 0) | \ 25 (folio_test_private(folio) ? PAGEMAP_BUFFERS : 0) \ 30 TP_PROTO(struct folio *folio), 32 TP_ARGS(folio), [all...] |
/linux-master/include/linux/ |
H A D | hugetlb_cgroup.h | 61 __hugetlb_cgroup_from_folio(struct folio *folio, bool rsvd) argument 63 VM_BUG_ON_FOLIO(!folio_test_hugetlb(folio), folio); 65 return folio->_hugetlb_cgroup_rsvd; 67 return folio->_hugetlb_cgroup; 70 static inline struct hugetlb_cgroup *hugetlb_cgroup_from_folio(struct folio *folio) argument 72 return __hugetlb_cgroup_from_folio(folio, false); 76 hugetlb_cgroup_from_folio_rsvd(struct folio *foli argument 81 __set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg, bool rsvd) argument 91 set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg) argument 97 set_hugetlb_cgroup_rsvd(struct folio *folio, struct hugetlb_cgroup *h_cg) argument 167 hugetlb_cgroup_from_folio(struct folio *folio) argument 173 hugetlb_cgroup_from_folio_rsvd(struct folio *folio) argument 178 set_hugetlb_cgroup(struct folio *folio, struct hugetlb_cgroup *h_cg) argument 183 set_hugetlb_cgroup_rsvd(struct folio *folio, struct hugetlb_cgroup *h_cg) argument 220 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 227 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 233 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument 238 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument [all...] |
H A D | cacheflush.h | 7 struct folio; 11 void flush_dcache_folio(struct folio *folio); 14 static inline void flush_dcache_folio(struct folio *folio) argument
|
H A D | rmap.h | 174 struct anon_vma *folio_get_anon_vma(struct folio *folio); 197 static inline void __folio_rmap_sanity_checks(struct folio *folio, argument 201 VM_WARN_ON_FOLIO(folio_test_hugetlb(folio), folio); local 213 VM_WARN_ON_FOLIO(page_folio(page) != folio, folio); local 214 VM_WARN_ON_FOLIO(page_folio(page + nr_pages - 1) != folio, folio); 225 VM_WARN_ON_FOLIO(folio_nr_pages(folio) != HPAGE_PMD_NR, folio); local 264 hugetlb_try_dup_anon_rmap(struct folio *folio, struct vm_area_struct *vma) argument 281 hugetlb_try_share_anon_rmap(struct folio *folio) argument 304 hugetlb_add_file_rmap(struct folio *folio) argument 307 VM_WARN_ON_FOLIO(folio_test_anon(folio), folio); local 313 hugetlb_remove_rmap(struct folio *folio) argument 321 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 357 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument 363 folio_dup_file_rmap_pte(struct folio *folio, struct page *page) argument 378 folio_dup_file_rmap_pmd(struct folio *folio, struct page *page) argument 388 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument 472 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument 479 folio_try_dup_anon_rmap_pte(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument 508 folio_try_dup_anon_rmap_pmd(struct folio *folio, struct page *page, struct vm_area_struct *src_vma) argument 520 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 613 folio_try_share_anon_rmap_pte(struct folio *folio, struct page *page) argument 642 folio_try_share_anon_rmap_pmd(struct folio *folio, struct page *page) argument 772 folio_referenced(struct folio *folio, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) argument 780 try_to_unmap(struct folio *folio, enum ttu_flags flags) argument 784 folio_mkclean(struct folio *folio) argument [all...] |
H A D | zswap.h | 30 bool zswap_store(struct folio *folio); 31 bool zswap_load(struct folio *folio); 37 void zswap_folio_swapin(struct folio *folio); 43 static inline bool zswap_store(struct folio *folio) argument 48 static inline bool zswap_load(struct folio *folio) argument 61 zswap_folio_swapin(struct folio *folio) argument [all...] |
H A D | page_idle.h | 14 static inline bool folio_test_young(const struct folio *folio) argument 16 struct page_ext *page_ext = page_ext_get(&folio->page); 28 static inline void folio_set_young(struct folio *folio) argument 30 struct page_ext *page_ext = page_ext_get(&folio->page); 39 static inline bool folio_test_clear_young(struct folio *folio) argument 41 struct page_ext *page_ext = page_ext_get(&folio->page); 53 static inline bool folio_test_idle(const struct folio *foli argument 67 folio_set_idle(struct folio *folio) argument 78 folio_clear_idle(struct folio *folio) argument [all...] |
H A D | migrate.h | 10 typedef struct folio *new_folio_t(struct folio *folio, unsigned long private); 11 typedef void free_folio_t(struct folio *folio, unsigned long private); 66 int migrate_folio_extra(struct address_space *mapping, struct folio *dst, 67 struct folio *src, enum migrate_mode mode, int extra_count); 68 int migrate_folio(struct address_space *mapping, struct folio *dst, 69 struct folio *src, enum migrate_mode mode); 73 struct folio *alloc_migration_targe 121 folio_test_movable(struct folio *folio) argument 127 folio_movable_ops(struct folio *folio) argument 148 migrate_misplaced_folio(struct folio *folio, struct vm_area_struct *vma, int node) argument [all...] |
H A D | page_ref.h | 71 * folio_ref_count - The reference count on this folio. 72 * @folio: The folio. 76 * folio refcount: 85 * Return: The number of references to this folio. 87 static inline int folio_ref_count(const struct folio *folio) argument 89 return page_ref_count(&folio->page); 104 static inline void folio_set_count(struct folio *folio, in argument 125 folio_ref_add(struct folio *folio, int nr) argument 137 folio_ref_sub(struct folio *folio, int nr) argument 142 folio_ref_sub_return(struct folio *folio, int nr) argument 158 folio_ref_inc(struct folio *folio) argument 170 folio_ref_dec(struct folio *folio) argument 184 folio_ref_sub_and_test(struct folio *folio, int nr) argument 198 folio_ref_inc_return(struct folio *folio) argument 212 folio_ref_dec_and_test(struct folio *folio) argument 226 folio_ref_dec_return(struct folio *folio) argument 240 folio_ref_add_unless(struct folio *folio, int nr, int u) argument 256 folio_try_get(struct folio *folio) argument 261 folio_ref_try_add_rcu(struct folio *folio, int count) argument 272 VM_BUG_ON_FOLIO(folio_ref_count(folio) == 0, folio); local 306 folio_try_get_rcu(struct folio *folio) argument 320 folio_ref_freeze(struct folio *folio, int count) argument 335 folio_ref_unfreeze(struct folio *folio, int count) argument [all...] |
H A D | mpage.h | 19 int mpage_read_folio(struct folio *folio, get_block_t get_block);
|
H A D | pagemap.h | 210 folio contents */ 366 * Large folio support currently depends on THP. These dependencies are 405 struct address_space *folio_mapping(struct folio *); 406 struct address_space *swapcache_mapping(struct folio *); 409 * folio_file_mapping - Find the mapping this folio belongs to. 410 * @folio: The folio. 420 static inline struct address_space *folio_file_mapping(struct folio *folio) argument 422 if (unlikely(folio_test_swapcache(folio))) 440 folio_flush_mapping(struct folio *folio) argument 462 folio_inode(struct folio *folio) argument 475 folio_attach_private(struct folio *folio, void *data) argument 493 folio_change_private(struct folio *folio, void *data) argument 510 folio_detach_private(struct folio *folio) argument 796 folio_index(struct folio *folio) argument 809 folio_next_index(struct folio *folio) argument 824 folio_file_page(struct folio *folio, pgoff_t index) argument 839 folio_contains(struct folio *folio, pgoff_t index) argument 932 folio_pos(struct folio *folio) argument 944 folio_file_pos(struct folio *folio) argument 952 folio_pgoff(struct folio *folio) argument 967 struct folio *folio; member in struct:wait_page_key 973 struct folio *folio; member in struct:wait_page_queue 1009 folio_trylock(struct folio *folio) argument 1044 folio_lock(struct folio *folio) argument 1064 struct folio *folio; local 1082 folio_lock_killable(struct folio *folio) argument 1097 folio_lock_or_retry(struct folio *folio, struct vm_fault *vmf) argument 1120 folio_wait_locked(struct folio *folio) argument 1126 folio_wait_locked_killable(struct folio *folio) argument 1149 folio_cancel_dirty(struct folio *folio) argument 1312 page_cache_async_readahead(struct address_space *mapping, struct file_ra_state *ra, struct file *file, struct folio *folio, pgoff_t index, unsigned long req_count) argument 1322 struct folio *folio; local 1351 struct folio *folio = __readahead_folio(ractl); local 1366 struct folio *folio = __readahead_folio(ractl); local 1475 folio_mkwrite_check_truncate(struct folio *folio, struct inode *inode) argument 1535 i_blocks_per_folio(struct inode *inode, struct folio *folio) argument [all...] |
/linux-master/fs/gfs2/ |
H A D | aops.h | 12 void gfs2_trans_add_databufs(struct gfs2_inode *ip, struct folio *folio,
|
/linux-master/fs/freevxfs/ |
H A D | vxfs_immed.c | 19 * @folio: folio to fill in. 23 * file that hosts @folio into the pagecache. 29 * @folio is locked and will be unlocked. 31 static int vxfs_immed_read_folio(struct file *fp, struct folio *folio) argument 33 struct vxfs_inode_info *vip = VXFS_INO(folio->mapping->host); 34 void *src = vip->vii_immed.vi_immed + folio_pos(folio); 37 for (i = 0; i < folio_nr_pages(folio); i++) { 38 memcpy_to_page(folio_page(folio, [all...] |
/linux-master/fs/bcachefs/ |
H A D | fs-io-pagecache.h | 7 typedef DARRAY(struct folio *) folios; 14 * Use u64 for the end pos and sector helpers because if the folio covers the 15 * max supported range of the mapping, the start offset of the next folio 19 static inline u64 folio_end_pos(struct folio *folio) argument 21 return folio_pos(folio) + folio_size(folio); 24 static inline size_t folio_sectors(struct folio *folio) argument 26 return PAGE_SECTORS << folio_order(folio); 29 folio_sector(struct folio *folio) argument 34 folio_end_sector(struct folio *folio) argument 72 bch2_folio_sector_set(struct folio *folio, struct bch_folio *s, unsigned i, unsigned n) argument 80 folio_pos_to_s(struct folio *folio, loff_t pos) argument 89 __bch2_folio_release(struct folio *folio) argument 94 bch2_folio_release(struct folio *folio) argument 100 __bch2_folio(struct folio *folio) argument 107 bch2_folio(struct folio *folio) argument [all...] |
/linux-master/fs/btrfs/ |
H A D | subpage.h | 10 struct folio; 46 * For locked bitmaps, normally it's subpage representation for folio 49 * - Metadata doesn't really lock the folio 95 struct folio *folio, enum btrfs_subpage_type type); 96 void btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio); 103 void btrfs_folio_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio); 104 void btrfs_folio_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *foli [all...] |
H A D | subpage.c | 121 struct folio *folio, enum btrfs_subpage_type type) 129 if (folio->mapping) 130 ASSERT(folio_test_locked(folio)); 132 /* Either not subpage, or the folio already has private attached. */ 133 if (!btrfs_is_subpage(fs_info, folio->mapping) || folio_test_private(folio)) 140 folio_attach_private(folio, subpage); 144 void btrfs_detach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio) argument 120 btrfs_attach_subpage(const struct btrfs_fs_info *fs_info, struct folio *folio, enum btrfs_subpage_type type) argument 195 btrfs_folio_inc_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio) argument 209 btrfs_folio_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct folio *folio) argument 224 btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 253 btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 275 btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 309 btrfs_subpage_clamp_range(struct folio *folio, u64 *start, u32 *len) argument 327 btrfs_subpage_start_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 347 btrfs_subpage_end_and_test_writer(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 390 btrfs_folio_start_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 407 btrfs_folio_end_writer_lock(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 429 btrfs_subpage_set_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 444 btrfs_subpage_clear_uptodate(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 458 btrfs_subpage_set_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 482 btrfs_subpage_clear_and_test_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 499 btrfs_subpage_clear_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 509 btrfs_subpage_set_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 524 btrfs_subpage_clear_writeback(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 541 btrfs_subpage_set_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 555 btrfs_subpage_clear_ordered(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 570 btrfs_subpage_set_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 585 btrfs_subpage_clear_checked(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 706 btrfs_folio_assert_not_dirty(const struct btrfs_fs_info *fs_info, struct folio *folio) argument 736 btrfs_folio_unlock_writer(struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument 772 btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info, struct folio *folio, u64 start, u32 len) argument [all...] |
/linux-master/fs/netfs/ |
H A D | misc.c | 12 * Attach a folio to the buffer and maybe set marks on it to say that we need 13 * to put the folio later and twiddle the pagecache flags. 16 struct folio *folio, unsigned int flags, 19 XA_STATE_ORDER(xas, xa, index, folio_order(folio)); 24 xas_store(&xas, folio); 50 struct folio *folio; local 57 /* TODO: Figure out what order folio can be allocated here */ 58 folio 15 netfs_xa_store_and_mark(struct xarray *xa, unsigned long index, struct folio *folio, unsigned int flags, gfp_t gfp_mask) argument 81 struct folio *folio; local 103 netfs_dirty_folio(struct address_space *mapping, struct folio *folio) argument 178 netfs_invalidate_folio(struct folio *folio, size_t offset, size_t length) argument 238 netfs_release_folio(struct folio *folio, gfp_t gfp) argument [all...] |
/linux-master/mm/damon/ |
H A D | ops-common.c | 22 struct folio *damon_get_folio(unsigned long pfn) 25 struct folio *folio; local 30 folio = page_folio(page); 31 if (!folio_test_lru(folio) || !folio_try_get(folio)) 33 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) { 34 folio_put(folio); 35 folio 42 struct folio *folio = damon_get_folio(pte_pfn(ptep_get(pte))); local 57 struct folio *folio = damon_get_folio(pmd_pfn(pmdp_get(pmd))); local [all...] |
H A D | paddr.c | 19 static bool damon_folio_mkold_one(struct folio *folio, argument 22 DEFINE_FOLIO_VMA_WALK(pvmw, folio, vma, addr, 0); 34 static void damon_folio_mkold(struct folio *folio) argument 42 if (!folio_mapped(folio) || !folio_raw_mapping(folio)) { 43 folio_set_idle(folio); 47 need_lock = !folio_test_anon(folio) || folio_test_ksm(folio); 60 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); local 87 damon_folio_young_one(struct folio *folio, struct vm_area_struct *vma, unsigned long addr, void *arg) argument 119 damon_folio_young(struct folio *folio) argument 150 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); local 198 __damos_pa_filter_out(struct damos_filter *filter, struct folio *folio) argument 232 damos_pa_filter_out(struct damos *scheme, struct folio *folio) argument 265 struct folio *folio = damon_get_folio(PHYS_PFN(addr)); local 297 struct folio *folio = damon_get_folio(PHYS_PFN(addr)); local [all...] |
/linux-master/fs/coda/ |
H A D | symlink.c | 23 static int coda_symlink_filler(struct file *file, struct folio *folio) argument 25 struct inode *inode = folio->mapping->host; 29 char *p = folio_address(folio); 36 folio_mark_uptodate(folio); 37 folio_unlock(folio); 41 folio_set_error(folio); 42 folio_unlock(folio);
|
/linux-master/fs/nilfs2/ |
H A D | page.c | 28 static struct buffer_head *__nilfs_get_folio_block(struct folio *folio, argument 34 struct buffer_head *bh = folio_buffers(folio); 37 bh = create_empty_buffers(folio, 1 << blkbits, b_state); 54 struct folio *folio; local 57 folio = filemap_grab_folio(mapping, index); 58 if (IS_ERR(folio)) 61 bh = __nilfs_get_folio_block(folio, blkoff, index, blkbits, b_state); 63 folio_unlock(folio); 76 struct folio *folio = bh->b_folio; local 140 nilfs_folio_buffers_clean(struct folio *folio) argument 153 nilfs_folio_bug(struct folio *folio) argument 258 struct folio *folio = fbatch.folios[i], *dfolio; local 312 struct folio *folio = fbatch.folios[i], *dfolio; local 373 struct folio *folio = fbatch.folios[i]; local 397 nilfs_clear_folio_dirty(struct folio *folio, bool silent) argument 462 __nilfs_clear_folio_dirty(struct folio *folio) argument 501 struct folio *folio; local [all...] |